[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/osism/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T01:37:29.194627Z",
                    "start": "2026-03-26T01:36:58.230743Z"
                },
                "id": "0242ac1f-6606-43e6-029b-000000000002",
                "name": "Base pre"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:36:58.282053Z",
                            "start": "2026-03-26T01:36:58.251085Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000000c",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "6d5/osism/6d507829b6994532b2cddf15505f7f09"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000087",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:36:58.329991Z",
                            "start": "2026-03-26T01:36:58.299653Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000089",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.14\nJob: testbed-upgrade-stable-rc-ubuntu-24.04\nPipeline: periodic-midnight\nExecutor: 521e9411259a\nTriggered by: https://github.com/osism/testbed\nEvent ID: 92c960495d4748b39c48d3d797ac3182\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:36:58.383870Z",
                            "start": "2026-03-26T01:36:58.341793Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000000d",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: orchestrator\nHostname: zuul-static-regiocloud-infra-1\nUsername: zuul-testbed03\nDistro: Debian 12.13\nProvider: static-testbed\nRegion: \nLabel: testbed-orchestrator\nProduct Name: OpenStack Nova\nInterface IP: 81.163.193.140\n",
                                    "zj_item": "orchestrator"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:36:58.532995Z",
                            "start": "2026-03-26T01:36:58.390797Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000000e",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:36:59.041211Z",
                            "start": "2026-03-26T01:36:58.554590Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000013",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "27b0a70a539c0ca8894f7a216997c337840a3b3b",
                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "27b0a70a539c0ca8894f7a216997c337840a3b3b",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489019.0916524-34-34228454365649/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "7335b2b501bd31b57ae7f963c71fdca7",
                            "mode": "0644",
                            "owner": "root",
                            "size": 7725,
                            "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489019.0916524-34-34228454365649/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:00.134126Z",
                            "start": "2026-03-26T01:36:59.057275Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000014",
                        "name": "Copy ansible inventory to logs dir"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:00.477595Z",
                            "start": "2026-03-26T01:37:00.157932Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:00.526602Z",
                            "start": "2026-03-26T01:37:00.490986Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:01.282964",
                            "end": "2026-03-26 01:37:02.295318",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-43e6-029b-0000000000aa-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:37:01.012354",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa.pub\nThe key fingerprint is:\nSHA256:xWyU5t0EaFjHT/UHZmymIt/WPO4ZhZRyIYacKf83A2s zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|         +oB=o=..|\n|        o+Xo.+*+.|\n|         *=..O+ o|\n|        .o+ ++o..|\n|        So + =. .|\n|          . E B. |\n|           o o.+ |\n|              .o |\n|             .o  |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:xWyU5t0EaFjHT/UHZmymIt/WPO4ZhZRyIYacKf83A2s zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|         +oB=o=..|",
                                "|        o+Xo.+*+.|",
                                "|         *=..O+ o|",
                                "|        .o+ ++o..|",
                                "|        So + =. .|",
                                "|          . E B. |",
                                "|           o o.+ |",
                                "|              .o |",
                                "|             .o  |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac1f-6606-43e6-029b-0000000000aa-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:02.345678Z",
                            "start": "2026-03-26T01:37:00.554672Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000aa",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:02.392853Z",
                            "start": "2026-03-26T01:37:02.354450Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000ab",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:02.439356Z",
                            "start": "2026-03-26T01:37:02.415463Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000ca",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiGosbjFN+VKLVAAW4MaKK1p284N8RWMRKkKqiULio+JY+rR819U6DjxCGg+fXIBh3XGz2z3gDc6WfZgG2AKcQKctNOf+fGww3c6RC6SxoXeKXZsO1Cu1c3iNH3xmavhWZgel4BbM7u1a3rTS/rs0XIYIEynby207yXEYURr13Z+0qBp3SH5VIadwoHjnalGWqHXcUDBnnnpSZacsS1s5IeDjmZabYyTq3bV7U/bdMo6kWZ2Lk9gmzqdFKgHW26n6AGTbqLPAGxWa6VpPFqNxmAxgP9wKDIHqPbjTTSjrgWsWBgNg91bYRoUm6nwmu7x2kBkRrfCP43MeylLqZeuurnqYiFOK3OEkzi+N4+PnCqE3F3hF/POX9Mu1tjjSZEffumApnUzB/h/6T9KdIr8maOoGP38ZyUko+vHcCnogNY3zeoidk9b3uVMlbPxeymXtTn5dV2UYXaXAwkybq1MGO+rNpk+hUTFrV61Ym3Pq/7ZJyCYSAkcbouuQ6eRHhNzs= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul-testbed03",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiGosbjFN+VKLVAAW4MaKK1p284N8RWMRKkKqiULio+JY+rR819U6DjxCGg+fXIBh3XGz2z3gDc6WfZgG2AKcQKctNOf+fGww3c6RC6SxoXeKXZsO1Cu1c3iNH3xmavhWZgel4BbM7u1a3rTS/rs0XIYIEynby207yXEYURr13Z+0qBp3SH5VIadwoHjnalGWqHXcUDBnnnpSZacsS1s5IeDjmZabYyTq3bV7U/bdMo6kWZ2Lk9gmzqdFKgHW26n6AGTbqLPAGxWa6VpPFqNxmAxgP9wKDIHqPbjTTSjrgWsWBgNg91bYRoUm6nwmu7x2kBkRrfCP43MeylLqZeuurnqYiFOK3OEkzi+N4+PnCqE3F3hF/POX9Mu1tjjSZEffumApnUzB/h/6T9KdIr8maOoGP38ZyUko+vHcCnogNY3zeoidk9b3uVMlbPxeymXtTn5dV2UYXaXAwkybq1MGO+rNpk+hUTFrV61Ym3Pq/7ZJyCYSAkcbouuQ6eRHhNzs= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul-testbed03",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:03.098674Z",
                            "start": "2026-03-26T01:37:02.449358Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000cb",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed03/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed03/.ssh"
                                }
                            },
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed03/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul-testbed03",
                            "path": "/home/zuul-testbed03/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:03.419808Z",
                            "start": "2026-03-26T01:37:03.108271Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000cc",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed03/.ssh/id_rsa",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa",
                                "force": false,
                                "mode": 384,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa",
                                    "force": false,
                                    "mode": 384,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa"
                                },
                                "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa"
                            },
                            "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:03.916135Z",
                            "start": "2026-03-26T01:37:03.428683Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000cd",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed03/.ssh/id_rsa.pub",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa.pub",
                                "force": false,
                                "mode": 420,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa.pub",
                                    "force": false,
                                    "mode": 420,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa.pub"
                                },
                                "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa.pub"
                            },
                            "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa.pub"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:04.396433Z",
                            "start": "2026-03-26T01:37:03.922137Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000ce",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:04.431046Z",
                            "start": "2026-03-26T01:37:04.406216Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000ac",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/var/ssh/nodepool"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-0000000000b1",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:04.904442Z",
                            "start": "2026-03-26T01:37:04.442026Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000b3",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa"
                            ],
                            "delta": "0:00:00.018296",
                            "end": "2026-03-26 01:37:05.266917",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-43e6-029b-0000000000b9-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:37:05.248621",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/6d507829b6994532b2cddf15505f7f09_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-43e6-029b-0000000000b9-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:05.310597Z",
                            "start": "2026-03-26T01:37:04.928508Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000b9",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:05.776256Z",
                            "start": "2026-03-26T01:37:05.322956Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000ba",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:05.819119Z",
                            "start": "2026-03-26T01:37:05.783779Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-0000000000bb",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001b",
                        "name": "start-zuul-console",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/start-zuul-console"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:06.301505Z",
                            "start": "2026-03-26T01:37:05.876703Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_info_dir": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:06.363981Z",
                            "start": "2026-03-26T01:37:06.316980Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000021",
                        "name": "Define zuul_info_dir fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:06.704643Z",
                            "start": "2026-03-26T01:37:06.374019Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000022",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "setup",
                            "ansible_facts": {
                                "ansible_all_ipv4_addresses": [
                                    "10.128.0.21"
                                ],
                                "ansible_all_ipv6_addresses": [
                                    "2a13:1a81:8000:2::33d",
                                    "fe80::f816:3eff:febf:60f0"
                                ],
                                "ansible_apparmor": {
                                    "status": "enabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "04/01/2014",
                                "ansible_bios_vendor": "SeaBIOS",
                                "ansible_bios_version": "1.15.0-1",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "QEMU",
                                "ansible_chassis_version": "pc-i440fx-6.2",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": "ttyS0,115200",
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_date_time": {
                                    "date": "2026-03-26",
                                    "day": "26",
                                    "epoch": "1774489027",
                                    "epoch_int": "1774489027",
                                    "hour": "01",
                                    "iso8601": "2026-03-26T01:37:07Z",
                                    "iso8601_basic": "20260326T013707484628",
                                    "iso8601_basic_short": "20260326T013707",
                                    "iso8601_micro": "2026-03-26T01:37:07.484628Z",
                                    "minute": "37",
                                    "month": "03",
                                    "second": "07",
                                    "time": "01:37:07",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Thursday",
                                    "weekday_number": "4",
                                    "weeknumber": "12",
                                    "year": "2026"
                                },
                                "ansible_default_ipv4": {
                                    "address": "10.128.0.21",
                                    "alias": "ens3",
                                    "broadcast": "",
                                    "gateway": "10.128.0.17",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "netmask": "255.255.255.240",
                                    "network": "10.128.0.16",
                                    "prefix": "28",
                                    "type": "ether"
                                },
                                "ansible_default_ipv6": {
                                    "address": "2a13:1a81:8000:2::33d",
                                    "gateway": "fe80::f816:3eff:fe98:d9f",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "prefix": "128",
                                    "scope": "global",
                                    "type": "ether"
                                },
                                "ansible_device_links": {
                                    "ids": {
                                        "sda": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                        ],
                                        "sda1": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                        ],
                                        "sda14": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                        ],
                                        "sda15": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                        ]
                                    },
                                    "labels": {},
                                    "masters": {},
                                    "uuids": {
                                        "sda1": [
                                            "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                        ],
                                        "sda15": [
                                            "1E24-44C9"
                                        ]
                                    }
                                },
                                "ansible_devices": {
                                    "loop0": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop1": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop2": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop3": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop4": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop5": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop6": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop7": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "sda": {
                                        "holders": [],
                                        "host": "SCSI storage controller: Red Hat, Inc. Virtio SCSI",
                                        "links": {
                                            "ids": [
                                                "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                            ],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": "QEMU HARDDISK",
                                        "partitions": {
                                            "sda1": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                                    ]
                                                },
                                                "sectors": "209453023",
                                                "sectorsize": 512,
                                                "size": "99.87 GB",
                                                "start": "262144",
                                                "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                            },
                                            "sda14": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": []
                                                },
                                                "sectors": "6144",
                                                "sectorsize": 512,
                                                "size": "3.00 MB",
                                                "start": "2048",
                                                "uuid": null
                                            },
                                            "sda15": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "1E24-44C9"
                                                    ]
                                                },
                                                "sectors": "253952",
                                                "sectorsize": 512,
                                                "size": "124.00 MB",
                                                "start": "8192",
                                                "uuid": "1E24-44C9"
                                            }
                                        },
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "209715200",
                                        "sectorsize": "512",
                                        "size": "100.00 GB",
                                        "support_discard": "4096",
                                        "vendor": "QEMU",
                                        "virtual": 1
                                    }
                                },
                                "ansible_distribution": "Debian",
                                "ansible_distribution_file_parsed": true,
                                "ansible_distribution_file_path": "/etc/os-release",
                                "ansible_distribution_file_variety": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_minor_version": "13",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12.13",
                                "ansible_dns": {
                                    "nameservers": [
                                        "81.163.194.9",
                                        "81.163.194.10",
                                        "2a13:1a81::11:53",
                                        "2a13:1a81::12:53"
                                    ],
                                    "search": [
                                        "."
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 1004,
                                "ansible_effective_user_id": 1003,
                                "ansible_ens3": {
                                    "active": true,
                                    "device": "ens3",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "off [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "off [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "on",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "on [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "off [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "off [fixed]",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "on [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "off [fixed]",
                                        "tx_nocache_copy": "off",
                                        "tx_scatter_gather": "on",
                                        "tx_scatter_gather_fraglist": "off [fixed]",
                                        "tx_sctp_segmentation": "off [fixed]",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "off",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "off [fixed]",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "off [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "10.128.0.21",
                                        "broadcast": "",
                                        "netmask": "255.255.255.240",
                                        "network": "10.128.0.16",
                                        "prefix": "28"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "2a13:1a81:8000:2::33d",
                                            "prefix": "128",
                                            "scope": "global"
                                        },
                                        {
                                            "address": "fe80::f816:3eff:febf:60f0",
                                            "prefix": "64",
                                            "scope": "link"
                                        }
                                    ],
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "module": "virtio_net",
                                    "mtu": 1500,
                                    "pciid": "virtio1",
                                    "promisc": false,
                                    "speed": -1,
                                    "timestamping": [],
                                    "type": "ether"
                                },
                                "ansible_env": {
                                    "HOME": "/home/zuul-testbed03",
                                    "LANG": "C.UTF-8",
                                    "LOGNAME": "zuul-testbed03",
                                    "MOTD_SHOWN": "pam",
                                    "PATH": "/usr/local/bin:/usr/bin:/bin:/usr/games",
                                    "PWD": "/home/zuul-testbed03",
                                    "SHELL": "/bin/sh",
                                    "SSH_CLIENT": "81.163.193.69 43598 22",
                                    "SSH_CONNECTION": "81.163.193.69 43598 10.128.0.21 22",
                                    "USER": "zuul-testbed03",
                                    "XDG_RUNTIME_DIR": "/run/user/1003",
                                    "XDG_SESSION_CLASS": "user",
                                    "XDG_SESSION_ID": "23207",
                                    "XDG_SESSION_TYPE": "tty"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "Other",
                                "ansible_fqdn": "zuul-static-regiocloud-infra-1",
                                "ansible_hostname": "zuul-static-regiocloud-infra-1",
                                "ansible_hostnqn": "",
                                "ansible_interfaces": [
                                    "lo",
                                    "ens3"
                                ],
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "6.1.0-23-cloud-amd64",
                                "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15)",
                                "ansible_lo": {
                                    "active": true,
                                    "device": "lo",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "on [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "on [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "off [fixed]",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "off [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on [fixed]",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "on [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "on",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "off [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "on [fixed]",
                                        "tx_nocache_copy": "off [fixed]",
                                        "tx_scatter_gather": "on [fixed]",
                                        "tx_scatter_gather_fraglist": "on [fixed]",
                                        "tx_sctp_segmentation": "on",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "on",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "on",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "on [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "127.0.0.1",
                                        "broadcast": "",
                                        "netmask": "255.0.0.0",
                                        "network": "127.0.0.0",
                                        "prefix": "8"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "::1",
                                            "prefix": "128",
                                            "scope": "host"
                                        }
                                    ],
                                    "mtu": 65536,
                                    "promisc": false,
                                    "timestamping": [],
                                    "type": "loopback"
                                },
                                "ansible_loadavg": {
                                    "15m": 0.2548828125,
                                    "1m": 0.81640625,
                                    "5m": 0.4501953125
                                },
                                "ansible_local": {},
                                "ansible_locally_reachable_ips": {
                                    "ipv4": [
                                        "10.128.0.21",
                                        "127.0.0.0/8",
                                        "127.0.0.1"
                                    ],
                                    "ipv6": [
                                        "::1",
                                        "2a13:1a81:8000:2::33d",
                                        "fe80::f816:3eff:febf:60f0"
                                    ]
                                },
                                "ansible_lsb": {
                                    "codename": "bookworm",
                                    "description": "Debian GNU/Linux 12 (bookworm)",
                                    "id": "Debian",
                                    "major_release": "12",
                                    "release": "12"
                                },
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_machine_id": "28f0fb0ddd3c48c38b3aa31603cb444d",
                                "ansible_memfree_mb": 426,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 7118,
                                        "used": 834
                                    },
                                    "real": {
                                        "free": 426,
                                        "total": 7952,
                                        "used": 7526
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 7952,
                                "ansible_mounts": [
                                    {
                                        "block_available": 17136243,
                                        "block_size": 4096,
                                        "block_total": 25759134,
                                        "block_used": 8622891,
                                        "device": "/dev/sda1",
                                        "fstype": "ext4",
                                        "inode_available": 5911342,
                                        "inode_total": 6545408,
                                        "inode_used": 634066,
                                        "mount": "/",
                                        "options": "rw,relatime,discard,errors=remount-ro",
                                        "size_available": 70190051328,
                                        "size_total": 105509412864,
                                        "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                    },
                                    {
                                        "block_available": 57419,
                                        "block_size": 2048,
                                        "block_total": 63339,
                                        "block_used": 5920,
                                        "device": "/dev/sda15",
                                        "fstype": "vfat",
                                        "inode_available": 0,
                                        "inode_total": 0,
                                        "inode_used": 0,
                                        "mount": "/boot/efi",
                                        "options": "rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=ascii,shortname=mixed,utf8,errors=remount-ro",
                                        "size_available": 117594112,
                                        "size_total": 129718272,
                                        "uuid": "1E24-44C9"
                                    }
                                ],
                                "ansible_nodename": "zuul-static-regiocloud-infra-1",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": [
                                        "tty0",
                                        "ttyS0,115200"
                                    ],
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_processor": [
                                    "0",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "1",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "2",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "3",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 4,
                                "ansible_processor_nproc": 4,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 4,
                                "ansible_product_name": "OpenStack Nova",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "26.2.1",
                                "ansible_python": {
                                    "executable": "/usr/bin/python3",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 2,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        2,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.2",
                                "ansible_real_group_id": 1004,
                                "ansible_real_user_id": 1003,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "systemd",
                                "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAOVugYBPOvEOPI9DA30PvhAEr88mJeaFmKN4pkfnBHV/1A1OlDKtX7lKWCBfEUilcd+tQFVKXm4KY+K9q4J5x5tpd7cL7LRDCFKN+py3+jc6U5J9EZ0o5cVG84z0c3XLEL6cCljG9Etx/0pXjjELRh1SGuUZweEx0143OSiKs/9DAAAAFQCT7R5MDTPEdFXQcZxA/fPRmtWFswAAAIBp8Iz2Njfcxfnh/mdxMztzkdNXulDt679+ghvl5588eCIDPa9xMtojivk53JaV6V4R+wsBN+MaKufzMOoVxuDkZcmw940H/1kcr3yOsJcZIkU1GRkJiNnSEIWJ6dVaXMh+wwbjw+8xNCA3+n3fARqUK906ICT5RsRrBe7W2K7dIAAAAIAL9qFr0vTidcXXhyWKN8I2iBsHSKfb6cjbseShw3ZeZE+kJ3IXz3z9oe0OwKp52GhvJrOXEkxST9YbtutbsVz5iS7lDpisvIWfFtvRRYaOtyvE8b4PWpHrk4s5tssqGDQ7y7zSqmRrj1sf5w0pwokPGmoSnPWUi0Vp0yhlysI/1w==",
                                "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss",
                                "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAuyO5R2SR4XT/Y65sItV7OQrUJXmGJ+//L9L/VugDohOWInFrkaZ6yFJMccVS7fmApekshaCGYnkIVdXV72hzg=",
                                "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256",
                                "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAILxLVCAcH99lPquKCT3PwHsvi+569sV28I73LxchEjeg",
                                "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519",
                                "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ3YX5Qhzq+HT8jTGGcX7rLsF3GCwObHFepfMmBZuD+gTfvdY5sYFSoXiT70rLPBjqETBTS0SQwmo8sd5AJI1vdgIqI5Sry8p52ng8tDg6C5CONt+xAaYOmE+ZDD/6naNX56b/gmgqBRQqPQHPmuptNxTlimdAcpoi+p9/UFBLTObMYmNL+5DTqw1+XECJ5j5kj5Ez40NT5w/1XnelEMcdw06WJZrzNEFt4Qpr2Zf+ybh0s+Xcn9u7d/gCbwXIPJvJnT0iSkiOzFvXJgOzwpvPtoi/5cV7g1LmgWsMLadT6gHof2poQaofC+1Bi0KiB2ANARMumvVBULx03qKvsqulsnlQAix2ffMraa3aMdUzK5aZo3hBx+H1uyEznUlyG9TUT5zhLIhNObetbpJgv6OJHiknBW6v/80FGD9t1pDiAwwBc/tJ8TyKpwyAdGMiEBIIVxTzx8QfPuApPJBJZXhMo2FBwRAEDEuuHbZu0znjhKKz4zlFT8TDrW+W7iThTvc=",
                                "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": [
                                    ""
                                ],
                                "ansible_system_capabilities_enforced": "True",
                                "ansible_system_vendor": "OpenStack Foundation",
                                "ansible_uptime_seconds": 49819003,
                                "ansible_user_dir": "/home/zuul-testbed03",
                                "ansible_user_gecos": "",
                                "ansible_user_gid": 1004,
                                "ansible_user_id": "zuul-testbed03",
                                "ansible_user_shell": "/bin/sh",
                                "ansible_user_uid": 1003,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "guest",
                                "ansible_virtualization_tech_guest": [
                                    "openstack"
                                ],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "openstack",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "fact_path": "/etc/ansible/facts.d",
                                    "filter": [],
                                    "gather_subset": [
                                        "all"
                                    ],
                                    "gather_timeout": 10
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:07.970238Z",
                            "start": "2026-03-26T01:37:06.713902Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000023",
                        "name": "Collect information about the host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "validate_host_hostname_sanitized": "orchestrator"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:08.043856Z",
                            "start": "2026-03-26T01:37:07.979200Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000024",
                        "name": "Sanitize hostname"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "5debb976826ac47b0694a793feb37e49654d6ea0",
                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info/host-info.orchestrator.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "host-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "5debb976826ac47b0694a793feb37e49654d6ea0",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info/host-info.orchestrator.yaml",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489028.0774062-238-221619651915831/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "949b6844f4fd6de81eba47cafa40e6b6",
                            "mode": "0644",
                            "owner": "root",
                            "size": 19729,
                            "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489028.0774062-238-221619651915831/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:08.637107Z",
                            "start": "2026-03-26T01:37:08.051674Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000025",
                        "name": "Write out all ansible variables/facts known for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_zuul_nolog_return": true,
                            "action": "zuul_debug_info",
                            "changed": false,
                            "df_i": "Filesystem      Inodes  IUsed   IFree IUse% Mounted on\nudev           1015018    312 1014706    1% /dev\ntmpfs          1017896    537 1017359    1% /run\n/dev/sda1      6545408 634066 5911342   10% /\ntmpfs          1017896     47 1017849    1% /dev/shm\ntmpfs          1017896      2 1017894    1% /run/lock\n/dev/sda15           0      0       0     - /boot/efi\ntmpfs           203579     17  203562    1% /run/user/1001\ntmpfs           203579     17  203562    1% /run/user/1005\ntmpfs           203579     17  203562    1% /run/user/1006\ntmpfs           203579     17  203562    1% /run/user/1003\n",
                            "df_m": "Filesystem     1M-blocks  Used Available Use% Mounted on\nudev                3965     0      3965   0% /dev\ntmpfs                796     1       795   1% /run\n/dev/sda1         100622 29558     66939  31% /\ntmpfs               3977     1      3976   1% /dev/shm\ntmpfs                  5     0         5   0% /run/lock\n/dev/sda15           124    12       113  10% /boot/efi\ntmpfs                796     0       796   0% /run/user/1001\ntmpfs                796     0       796   0% /run/user/1005\ntmpfs                796     0       796   0% /run/user/1006\ntmpfs                796     0       796   0% /run/user/1003\n",
                            "image_manifest_files": [],
                            "invocation": {
                                "module_args": {
                                    "image_manifest": null,
                                    "image_manifest_files": [
                                        "/etc/dib-builddate.txt",
                                        "/etc/image-hostname.txt"
                                    ],
                                    "ipv4_route_required": false,
                                    "ipv6_route_required": false,
                                    "traceroute_host": null
                                }
                            },
                            "network_interfaces": "1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n    inet 127.0.0.1/8 scope host lo\n       valid_lft forever preferred_lft forever\n    inet6 ::1/128 scope host noprefixroute \n       valid_lft forever preferred_lft forever\n2: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000\n    link/ether fa:16:3e:bf:60:f0 brd ff:ff:ff:ff:ff:ff\n    altname enp0s3\n    inet 10.128.0.21/28 metric 100 brd 10.128.0.31 scope global dynamic ens3\n       valid_lft 40637sec preferred_lft 40637sec\n    inet6 2a13:1a81:8000:2::33d/128 scope global noprefixroute \n       valid_lft forever preferred_lft forever\n    inet6 fe80::f816:3eff:febf:60f0/64 scope link \n       valid_lft forever preferred_lft forever\n",
                            "network_neighbors": "10.128.0.17 dev ens3 lladdr fa:16:3e:e3:0c:52 REACHABLE \nfe80::f816:3eff:fe98:d9f dev ens3 lladdr fa:16:3e:98:0d:9f router STALE \n",
                            "network_routing_v4": "default via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n10.128.0.16/28 dev ens3 proto kernel scope link src 10.128.0.21 metric 100 \n10.128.0.17 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n10.128.0.18 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n81.163.194.9 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n81.163.194.10 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n169.254.169.254 via 10.128.0.18 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n",
                            "network_routing_v6": "2a13:1a81:8000:2::300/120 dev ens3 proto ra metric 100 pref medium\nfe80::/64 dev ens3 proto kernel metric 256 pref medium\ndefault via fe80::f816:3eff:fe98:d9f dev ens3 proto ra metric 100 mtu 1500 pref medium\n",
                            "proc_cpuinfo": "processor\t: 0\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 0\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 0\ninitial apicid\t: 0\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 1\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 1\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 1\ninitial apicid\t: 1\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 2\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 2\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 2\ninitial apicid\t: 2\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 3\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 3\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 3\ninitial apicid\t: 3\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\n",
                            "traceroute": null,
                            "uname": "Linux zuul-static-regiocloud-infra-1 6.1.0-23-cloud-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:09.111584Z",
                            "start": "2026-03-26T01:37:08.648529Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000027",
                        "name": "Collect information about zuul worker"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "b470d27e1aac3fce6df468fbfee3db45bbd80344",
                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info/zuul-info.orchestrator.txt",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "zuul-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "b470d27e1aac3fce6df468fbfee3db45bbd80344",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/zuul-info/zuul-info.orchestrator.txt",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489029.160589-277-241178233013818/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "e4e9ae487a183e99d4c6d9adcabb425a",
                            "mode": "0644",
                            "owner": "root",
                            "size": 3369,
                            "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/.ansible/tmp/ansible-tmp-1774489029.160589-277-241178233013818/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:09.679317Z",
                            "start": "2026-03-26T01:37:09.119567Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-000000000028",
                        "name": "Write out all zuul information for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:10.032442Z",
                            "start": "2026-03-26T01:37:09.698201Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000002c",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --compress --delete-after --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --out-format='<<CHANGED>>%i %n%L' /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/src zuul-testbed03@81.163.193.140:.",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": false,
                                    "delete": true,
                                    "dest": "zuul-testbed03@81.163.193.140:.",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": false,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "push",
                                    "owner": false,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": true,
                                    "rsync_opts": [],
                                    "rsync_path": null,
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/src",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": ".d..t...... src/\n.d..t...... src/github.com/\n.d..t...... src/github.com/osism/\n.d..t...... src/github.com/osism/ansible-collection-commons/\n<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-commons/.flake8\n<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-commons/CHANGELOG.md\n<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-commons/README.md\n<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/description\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/index\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/0b/1a4cc89d202712cb431174fd6158e7682f3b64\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/1a/e6b58d2ab820fda6fc7e19c31f98d3e0c226d4\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/47/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/a3/7092a7e1e69e3523f5770c7c02f4482413eed5\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/e5/8170b7f562ceaf1fbf5908015364fd9de2110d\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/\n<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-fast.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-fast.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-services/\n<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-services/.flake8\n<f..t...... src/github.com/osism/ansible-collection-services/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-services/CHANGELOG.md\n<f..t...... src/github.com/osism/ansible-collection-services/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-services/README.md\n<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-services/.git/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD\n<f.st...... src/github.com/osism/ansible-collection-services/.git/config\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-services/.git/description\n<f..t...... src/github.com/osism/ansible-collection-services/.git/index\n<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-services/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-services/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/56/14ee480009ffbd1fe623254aed3aaea46befc8\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/0e94d45a814c75065d42979425a0407f071c3a\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6f/2f892b7b18b7b44e432e38593a701da8667311\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/d2/71a5b882399ba36bbb36181be297a1fc53845c\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.github/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-services/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-services/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_metalbox.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_network.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2\n.d..t...... src/github.com/osism/terraform-base/\n<f..t...... src/github.com/osism/terraform-base/.yamllint.yml\n<f..t...... src/github.com/osism/terraform-base/.zuul.yaml\n<f..t...... src/github.com/osism/terraform-base/LICENSE\n<f..t...... src/github.com/osism/terraform-base/README.md\n.d..t...... src/github.com/osism/terraform-base/.git/\n<f..t...... src/github.com/osism/terraform-base/.git/HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/config\n<f..t...... src/github.com/osism/terraform-base/.git/config.worktree\n<f..t...... src/github.com/osism/terraform-base/.git/description\n<f..t...... src/github.com/osism/terraform-base/.git/index\n<f..t...... src/github.com/osism/terraform-base/.git/packed-refs\n.d..t...... src/github.com/osism/terraform-base/.git/branches/\n.d..t...... src/github.com/osism/terraform-base/.git/hooks/\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample\n.d..t...... src/github.com/osism/terraform-base/.git/info/\n<f..t...... src/github.com/osism/terraform-base/.git/info/exclude\n.d..t...... src/github.com/osism/terraform-base/.git/logs/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/objects/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/00/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/01/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/02/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/05/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/06/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/07/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/08/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/10/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/12/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/15/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/16/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/17/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/20/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/21/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/24/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/26/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/29/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/2c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/36/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/37/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/41/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/43/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/46/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/47/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/48/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/49/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/52/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/55/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/56/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/58/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/5d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/61/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/63/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/66/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/67/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/68/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/6d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/70/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/72/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/76/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/80/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/81/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/82/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/83/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/84/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/85/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/86/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/88/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/90/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/91/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/92/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/95/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/96/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/98/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9e/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/aa/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/af/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ba/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ce/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/da/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/db/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dd/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/de/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/df/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ea/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ee/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ef/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/fb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ff/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/info/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/pack/\n.d..t...... src/github.com/osism/terraform-base/.git/refs/\n.d..t...... src/github.com/osism/terraform-base/.github/\n<f..t...... src/github.com/osism/terraform-base/.github/renovate.json\n.d..t...... src/github.com/osism/terraform-base/playbooks/\n<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml\n.d..t...... src/github.com/osism/terraform-base/testbed-default/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/environments/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars\n.d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf\n.d..t...... src/github.com/osism/testbed/\n<f..t...... src/github.com/osism/testbed/.ansible-lint\n<f..t...... src/github.com/osism/testbed/.flake8\n<f..t...... src/github.com/osism/testbed/.gitignore\n<f..t...... src/github.com/osism/testbed/.yamllint.yml\n<f..t...... src/github.com/osism/testbed/.zuul.yaml\n<f..t...... src/github.com/osism/testbed/LICENSE\n<f..t...... src/github.com/osism/testbed/Makefile\n<f..t...... src/github.com/osism/testbed/README.md\n<f..t...... src/github.com/osism/testbed/ansible.cfg\n<f..t...... src/github.com/osism/testbed/gilt.yml\n<f..t...... src/github.com/osism/testbed/requirements.txt\n<f..t...... src/github.com/osism/testbed/requirements.yml\n<f..t...... src/github.com/osism/testbed/tox.ini\n.d..t...... src/github.com/osism/testbed/.git/\n<f..t...... src/github.com/osism/testbed/.git/HEAD\n<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/testbed/.git/config\n<f..t...... src/github.com/osism/testbed/.git/config.worktree\n<f..t...... src/github.com/osism/testbed/.git/description\n<f.st...... src/github.com/osism/testbed/.git/index\n<f..t...... src/github.com/osism/testbed/.git/packed-refs\n.d..t...... src/github.com/osism/testbed/.git/branches/\n.d..t...... src/github.com/osism/testbed/.git/hooks/\n<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample\n.d..t...... src/github.com/osism/testbed/.git/info/\n<f..t...... src/github.com/osism/testbed/.git/info/exclude\n.d..t...... src/github.com/osism/testbed/.git/logs/\n<f..t...... src/github.com/osism/testbed/.git/logs/HEAD\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/testbed/.git/objects/\n.d..t...... src/github.com/osism/testbed/.git/objects/00/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/00/30c229a586597810f7c072e6e96cc8fec6abd8\n.d..t...... src/github.com/osism/testbed/.git/objects/01/\n.d..t...... src/github.com/osism/testbed/.git/objects/02/\n.d..t...... src/github.com/osism/testbed/.git/objects/03/\n.d..t...... src/github.com/osism/testbed/.git/objects/04/\n.d..t...... src/github.com/osism/testbed/.git/objects/05/\n.d..t...... src/github.com/osism/testbed/.git/objects/06/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/06/c76461b7ccc166f01b2d33a4628084bb085d9d\n.d..t...... src/github.com/osism/testbed/.git/objects/07/\n.d..t...... src/github.com/osism/testbed/.git/objects/08/\n.d..t...... src/github.com/osism/testbed/.git/objects/09/\n.d..t...... src/github.com/osism/testbed/.git/objects/0a/\n.d..t...... src/github.com/osism/testbed/.git/objects/0b/\n.d..t...... src/github.com/osism/testbed/.git/objects/0c/\n.d..t...... src/github.com/osism/testbed/.git/objects/0d/\n.d..t...... src/github.com/osism/testbed/.git/objects/0e/\n.d..t...... src/github.com/osism/testbed/.git/objects/0f/\n.d..t...... src/github.com/osism/testbed/.git/objects/10/\n.d..t...... src/github.com/osism/testbed/.git/objects/11/\n.d..t...... src/github.com/osism/testbed/.git/objects/12/\n.d..t...... src/github.com/osism/testbed/.git/objects/13/\n.d..t...... src/github.com/osism/testbed/.git/objects/14/\n.d..t...... src/github.com/osism/testbed/.git/objects/15/\n.d..t...... src/github.com/osism/testbed/.git/objects/16/\n.d..t...... src/github.com/osism/testbed/.git/objects/17/\n.d..t...... src/github.com/osism/testbed/.git/objects/18/\n.d..t...... src/github.com/osism/testbed/.git/objects/19/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/19/c1e98eff29413b62f6534ae2c146bf36f85675\n.d..t...... src/github.com/osism/testbed/.git/objects/1a/\n.d..t...... src/github.com/osism/testbed/.git/objects/1b/\n.d..t...... src/github.com/osism/testbed/.git/objects/1c/\n.d..t...... src/github.com/osism/testbed/.git/objects/1d/\n.d..t...... src/github.com/osism/testbed/.git/objects/1e/\n.d..t...... src/github.com/osism/testbed/.git/objects/1f/\n.d..t...... src/github.com/osism/testbed/.git/objects/20/\n.d..t...... src/github.com/osism/testbed/.git/objects/21/\n.d..t...... src/github.com/osism/testbed/.git/objects/22/\n.d..t...... src/github.com/osism/testbed/.git/objects/23/\n.d..t...... src/github.com/osism/testbed/.git/objects/24/\n.d..t...... src/github.com/osism/testbed/.git/objects/25/\n.d..t...... src/github.com/osism/testbed/.git/objects/26/\n.d..t...... src/github.com/osism/testbed/.git/objects/27/\n.d..t...... src/github.com/osism/testbed/.git/objects/28/\n.d..t...... src/github.com/osism/testbed/.git/objects/29/\n.d..t...... src/github.com/osism/testbed/.git/objects/2a/\n.d..t...... src/github.com/osism/testbed/.git/objects/2b/\n.d..t...... src/github.com/osism/testbed/.git/objects/2c/\n.d..t...... src/github.com/osism/testbed/.git/objects/2d/\n.d..t...... src/github.com/osism/testbed/.git/objects/2e/\n.d..t...... src/github.com/osism/testbed/.git/objects/2f/\n.d..t...... src/github.com/osism/testbed/.git/objects/30/\n.d..t...... src/github.com/osism/testbed/.git/objects/31/\n.d..t...... src/github.com/osism/testbed/.git/objects/32/\n.d..t...... src/github.com/osism/testbed/.git/objects/33/\n.d..t...... src/github.com/osism/testbed/.git/objects/34/\n.d..t...... src/github.com/osism/testbed/.git/objects/35/\n.d..t...... src/github.com/osism/testbed/.git/objects/36/\n.d..t...... src/github.com/osism/testbed/.git/objects/37/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/37/c6b001771fb59f1ffee2c9dc64557ef410bdf0\n.d..t...... src/github.com/osism/testbed/.git/objects/38/\n.d..t...... src/github.com/osism/testbed/.git/objects/39/\n.d..t...... src/github.com/osism/testbed/.git/objects/3a/\n.d..t...... src/github.com/osism/testbed/.git/objects/3b/\n.d..t...... src/github.com/osism/testbed/.git/objects/3c/\n.d..t...... src/github.com/osism/testbed/.git/objects/3d/\n.d..t...... src/github.com/osism/testbed/.git/objects/3e/\n.d..t...... src/github.com/osism/testbed/.git/objects/3f/\n.d..t...... src/github.com/osism/testbed/.git/objects/40/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/40/4a99b95238a5949445182b00d9725dc2ecd54b\n.d..t...... src/github.com/osism/testbed/.git/objects/41/\n.d..t...... src/github.com/osism/testbed/.git/objects/42/\n.d..t...... src/github.com/osism/testbed/.git/objects/43/\n.d..t...... src/github.com/osism/testbed/.git/objects/44/\n.d..t...... src/github.com/osism/testbed/.git/objects/45/\n.d..t...... src/github.com/osism/testbed/.git/objects/46/\n.d..t...... src/github.com/osism/testbed/.git/objects/47/\n.d..t...... src/github.com/osism/testbed/.git/objects/48/\n.d..t...... src/github.com/osism/testbed/.git/objects/49/\n.d..t...... src/github.com/osism/testbed/.git/objects/4a/\n.d..t...... src/github.com/osism/testbed/.git/objects/4c/\n.d..t...... src/github.com/osism/testbed/.git/objects/4d/\n.d..t...... src/github.com/osism/testbed/.git/objects/4e/\n.d..t...... src/github.com/osism/testbed/.git/objects/4f/\n.d..t...... src/github.com/osism/testbed/.git/objects/50/\n.d..t...... src/github.com/osism/testbed/.git/objects/51/\n.d..t...... src/github.com/osism/testbed/.git/objects/52/\n.d..t...... src/github.com/osism/testbed/.git/objects/53/\n.d..t...... src/github.com/osism/testbed/.git/objects/54/\n.d..t...... src/github.com/osism/testbed/.git/objects/55/\n.d..t...... src/github.com/osism/testbed/.git/objects/56/\n.d..t...... src/github.com/osism/testbed/.git/objects/57/\n.d..t...... src/github.com/osism/testbed/.git/objects/58/\n.d..t...... src/github.com/osism/testbed/.git/objects/59/\n.d..t...... src/github.com/osism/testbed/.git/objects/5a/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/5a/0516c9d464c12ecd7c4379ea40b1f1ad5e485d\n.d..t...... src/github.com/osism/testbed/.git/objects/5b/\n.d..t...... src/github.com/osism/testbed/.git/objects/5c/\n.d..t...... src/github.com/osism/testbed/.git/objects/5d/\n.d..t...... src/github.com/osism/testbed/.git/objects/5e/\n.d..t...... src/github.com/osism/testbed/.git/objects/5f/\n.d..t...... src/github.com/osism/testbed/.git/objects/60/\n.d..t...... src/github.com/osism/testbed/.git/objects/61/\n.d..t...... src/github.com/osism/testbed/.git/objects/64/\n.d..t...... src/github.com/osism/testbed/.git/objects/65/\n.d..t...... src/github.com/osism/testbed/.git/objects/66/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/66/4d846e525baf80e14effbeb2b02e33d9174a5e\n.d..t...... src/github.com/osism/testbed/.git/objects/67/\n.d..t...... src/github.com/osism/testbed/.git/objects/68/\n.d..t...... src/github.com/osism/testbed/.git/objects/69/\n.d..t...... src/github.com/osism/testbed/.git/objects/6a/\n.d..t...... src/github.com/osism/testbed/.git/objects/6b/\n.d..t...... src/github.com/osism/testbed/.git/objects/6c/\n.d..t...... src/github.com/osism/testbed/.git/objects/6d/\n.d..t...... src/github.com/osism/testbed/.git/objects/6e/\n.d..t...... src/github.com/osism/testbed/.git/objects/6f/\n.d..t...... src/github.com/osism/testbed/.git/objects/70/\n.d..t...... src/github.com/osism/testbed/.git/objects/71/\n.d..t...... src/github.com/osism/testbed/.git/objects/72/\n.d..t...... src/github.com/osism/testbed/.git/objects/73/\n.d..t...... src/github.com/osism/testbed/.git/objects/74/\n.d..t...... src/github.com/osism/testbed/.git/objects/75/\n.d..t...... src/github.com/osism/testbed/.git/objects/76/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/76/a4c480e5f6e4bb27df8bda37f74d6388c4321b\n.d..t...... src/github.com/osism/testbed/.git/objects/77/\n.d..t...... src/github.com/osism/testbed/.git/objects/78/\n.d..t...... src/github.com/osism/testbed/.git/objects/79/\n.d..t...... src/github.com/osism/testbed/.git/objects/7a/\n.d..t...... src/github.com/osism/testbed/.git/objects/7b/\n.d..t...... src/github.com/osism/testbed/.git/objects/7c/\n.d..t...... src/github.com/osism/testbed/.git/objects/7d/\n.d..t...... src/github.com/osism/testbed/.git/objects/7e/\n.d..t...... src/github.com/osism/testbed/.git/objects/7f/\n.d..t...... src/github.com/osism/testbed/.git/objects/80/\n.d..t...... src/github.com/osism/testbed/.git/objects/81/\n.d..t...... src/github.com/osism/testbed/.git/objects/82/\n.d..t...... src/github.com/osism/testbed/.git/objects/83/\n.d..t...... src/github.com/osism/testbed/.git/objects/84/\n.d..t...... src/github.com/osism/testbed/.git/objects/85/\n.d..t...... src/github.com/osism/testbed/.git/objects/86/\n.d..t...... src/github.com/osism/testbed/.git/objects/87/\n.d..t...... src/github.com/osism/testbed/.git/objects/8a/\n.d..t...... src/github.com/osism/testbed/.git/objects/8b/\n.d..t...... src/github.com/osism/testbed/.git/objects/8c/\n.d..t...... src/github.com/osism/testbed/.git/objects/8e/\n.d..t...... src/github.com/osism/testbed/.git/objects/8f/\n.d..t...... src/github.com/osism/testbed/.git/objects/90/\n.d..t...... src/github.com/osism/testbed/.git/objects/91/\n.d..t...... src/github.com/osism/testbed/.git/objects/92/\n.d..t...... src/github.com/osism/testbed/.git/objects/93/\n.d..t...... src/github.com/osism/testbed/.git/objects/94/\n.d..t...... src/github.com/osism/testbed/.git/objects/95/\n.d..t...... src/github.com/osism/testbed/.git/objects/96/\n.d..t...... src/github.com/osism/testbed/.git/objects/97/\n.d..t...... src/github.com/osism/testbed/.git/objects/98/\n.d..t...... src/github.com/osism/testbed/.git/objects/9a/\n.d..t...... src/github.com/osism/testbed/.git/objects/9c/\n.d..t...... src/github.com/osism/testbed/.git/objects/9e/\n.d..t...... src/github.com/osism/testbed/.git/objects/9f/\n.d..t...... src/github.com/osism/testbed/.git/objects/a0/\n.d..t...... src/github.com/osism/testbed/.git/objects/a1/\n.d..t...... src/github.com/osism/testbed/.git/objects/a2/\n.d..t...... src/github.com/osism/testbed/.git/objects/a3/\n.d..t...... src/github.com/osism/testbed/.git/objects/a4/\n.d..t...... src/github.com/osism/testbed/.git/objects/a5/\n.d..t...... src/github.com/osism/testbed/.git/objects/a6/\n.d..t...... src/github.com/osism/testbed/.git/objects/a7/\n.d..t...... src/github.com/osism/testbed/.git/objects/a8/\n.d..t...... src/github.com/osism/testbed/.git/objects/a9/\n.d..t...... src/github.com/osism/testbed/.git/objects/aa/\n.d..t...... src/github.com/osism/testbed/.git/objects/ab/\n.d..t...... src/github.com/osism/testbed/.git/objects/ac/\n.d..t...... src/github.com/osism/testbed/.git/objects/ad/\n.d..t...... src/github.com/osism/testbed/.git/objects/ae/\n.d..t...... src/github.com/osism/testbed/.git/objects/af/\n.d..t...... src/github.com/osism/testbed/.git/objects/b0/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/b0/a678c098711df8f3fa53ff0af127883997a1a2\n.d..t...... src/github.com/osism/testbed/.git/objects/b1/\n.d..t...... src/github.com/osism/testbed/.git/objects/b2/\n.d..t...... src/github.com/osism/testbed/.git/objects/b3/\n.d..t...... src/github.com/osism/testbed/.git/objects/b4/\n.d..t...... src/github.com/osism/testbed/.git/objects/b5/\n.d..t...... src/github.com/osism/testbed/.git/objects/b6/\n.d..t...... src/github.com/osism/testbed/.git/objects/b7/\n.d..t...... src/github.com/osism/testbed/.git/objects/b8/\n.d..t...... src/github.com/osism/testbed/.git/objects/b9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ba/\n.d..t...... src/github.com/osism/testbed/.git/objects/bb/\n.d..t...... src/github.com/osism/testbed/.git/objects/bc/\n.d..t...... src/github.com/osism/testbed/.git/objects/bd/\n.d..t...... src/github.com/osism/testbed/.git/objects/be/\n.d..t...... src/github.com/osism/testbed/.git/objects/bf/\n.d..t...... src/github.com/osism/testbed/.git/objects/c0/\n.d..t...... src/github.com/osism/testbed/.git/objects/c1/\n.d..t...... src/github.com/osism/testbed/.git/objects/c2/\n.d..t...... src/github.com/osism/testbed/.git/objects/c3/\n.d..t...... src/github.com/osism/testbed/.git/objects/c4/\n.d..t...... src/github.com/osism/testbed/.git/objects/c5/\n.d..t...... src/github.com/osism/testbed/.git/objects/c6/\n.d..t...... src/github.com/osism/testbed/.git/objects/c7/\n.d..t...... src/github.com/osism/testbed/.git/objects/c8/\n.d..t...... src/github.com/osism/testbed/.git/objects/c9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ca/\n.d..t...... src/github.com/osism/testbed/.git/objects/cb/\n.d..t...... src/github.com/osism/testbed/.git/objects/cc/\n.d..t...... src/github.com/osism/testbed/.git/objects/cd/\n.d..t...... src/github.com/osism/testbed/.git/objects/ce/\n.d..t...... src/github.com/osism/testbed/.git/objects/cf/\n.d..t...... src/github.com/osism/testbed/.git/objects/d0/\n.d..t...... src/github.com/osism/testbed/.git/objects/d2/\n.d..t...... src/github.com/osism/testbed/.git/objects/d3/\n.d..t...... src/github.com/osism/testbed/.git/objects/d5/\n.d..t...... src/github.com/osism/testbed/.git/objects/d6/\n.d..t...... src/github.com/osism/testbed/.git/objects/d7/\n.d..t...... src/github.com/osism/testbed/.git/objects/d8/\n.d..t...... src/github.com/osism/testbed/.git/objects/d9/\n.d..t...... src/github.com/osism/testbed/.git/objects/da/\n.d..t...... src/github.com/osism/testbed/.git/objects/db/\n.d..t...... src/github.com/osism/testbed/.git/objects/dc/\n.d..t...... src/github.com/osism/testbed/.git/objects/dd/\n.d..t...... src/github.com/osism/testbed/.git/objects/de/\n.d..t...... src/github.com/osism/testbed/.git/objects/df/\n.d..t...... src/github.com/osism/testbed/.git/objects/e0/\n.d..t...... src/github.com/osism/testbed/.git/objects/e2/\n.d..t...... src/github.com/osism/testbed/.git/objects/e3/\n.d..t...... src/github.com/osism/testbed/.git/objects/e4/\n.d..t...... src/github.com/osism/testbed/.git/objects/e5/\n.d..t...... src/github.com/osism/testbed/.git/objects/e6/\n.d..t...... src/github.com/osism/testbed/.git/objects/e7/\n.d..t...... src/github.com/osism/testbed/.git/objects/e8/\n.d..t...... src/github.com/osism/testbed/.git/objects/e9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ea/\n.d..t...... src/github.com/osism/testbed/.git/objects/eb/\n.d..t...... src/github.com/osism/testbed/.git/objects/ec/\n.d..t...... src/github.com/osism/testbed/.git/objects/ed/\n.d..t...... src/github.com/osism/testbed/.git/objects/ee/\n.d..t...... src/github.com/osism/testbed/.git/objects/ef/\n.d..t...... src/github.com/osism/testbed/.git/objects/f0/\n.d..t...... src/github.com/osism/testbed/.git/objects/f1/\n.d..t...... src/github.com/osism/testbed/.git/objects/f2/\n.d..t...... src/github.com/osism/testbed/.git/objects/f3/\n.d..t...... src/github.com/osism/testbed/.git/objects/f4/\n.d..t...... src/github.com/osism/testbed/.git/objects/f5/\n.d..t...... src/github.com/osism/testbed/.git/objects/f6/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/f6/7fac5930587796a00257dbd2bf7f6faab966e3\n.d..t...... src/github.com/osism/testbed/.git/objects/f8/\n.d..t...... src/github.com/osism/testbed/.git/objects/f9/\n.d..t...... src/github.com/osism/testbed/.git/objects/fa/\n.d..t...... src/github.com/osism/testbed/.git/objects/fb/\n.d..t...... src/github.com/osism/testbed/.git/objects/fc/\n.d..t...... src/github.com/osism/testbed/.git/objects/fd/\n.d..t...... src/github.com/osism/testbed/.git/objects/fe/\n.d..t...... src/github.com/osism/testbed/.git/objects/ff/\n.d..t...... src/github.com/osism/testbed/.git/objects/info/\n.d..t...... src/github.com/osism/testbed/.git/objects/pack/\n.d..t...... src/github.com/osism/testbed/.git/refs/\n.d..t...... src/github.com/osism/testbed/.github/\n<f..t...... src/github.com/osism/testbed/.github/renovate.json\n.d..t...... src/github.com/osism/testbed/.github/workflows/\n<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml\n.d..t...... src/github.com/osism/testbed/ansible/\n<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml\n<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml\n.d..t...... src/github.com/osism/testbed/containerlab/\n<f..t...... src/github.com/osism/testbed/containerlab/README.md\n<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml\n.d..t...... src/github.com/osism/testbed/containerlab/files/\n<f..t...... src/github.com/osism/testbed/containerlab/files/daemons\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf\n.d..t...... src/github.com/osism/testbed/contrib/\n<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh\n<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py\n<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql\n<f..t...... src/github.com/osism/testbed/contrib/semver2.sh\n<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py\n.d..t...... src/github.com/osism/testbed/contrib/ansible/\n<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml\n<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample\n.d..t...... src/github.com/osism/testbed/contrib/assets/\n<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot\n.d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster\n.d..t...... src/github.com/osism/testbed/contrib/openstack-cli/\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example\n.d..t...... src/github.com/osism/testbed/contrib/ownca/\n<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml\n.d..t...... src/github.com/osism/testbed/environments/\n<f..t...... src/github.com/osism/testbed/environments/.vault_pass\n<f..t...... src/github.com/osism/testbed/environments/ansible.cfg\n<f..t...... src/github.com/osism/testbed/environments/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/images.yml\n<f..t...... src/github.com/osism/testbed/environments/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph.test/\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph/\n<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/\n<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/files/\n<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact\n.d..t...... src/github.com/osism/testbed/environments/custom/templates/\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2\n.d..t...... src/github.com/osism/testbed/environments/generic/\n<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/images.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/\n<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf\n<f+++++++++ src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/policy.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/files/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/\n<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/hosts\n<f..t...... src/github.com/osism/testbed/environments/manager/images.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/files/\n<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml\n.d..t...... src/github.com/osism/testbed/environments/monitoring/\n<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/openstack/\n<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub\n<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/openrc\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst\n.d..t...... src/github.com/osism/testbed/inventory/\n<f..t...... src/github.com/osism/testbed/inventory/10-custom\n<f..t...... src/github.com/osism/testbed/inventory/20-roles\n<f..t...... src/github.com/osism/testbed/inventory/99-overwrite\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/all/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml\n.d..t...... src/github.com/osism/testbed/netbox/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml\n.d..t...... src/github.com/osism/testbed/netbox/moduletypes/\n<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep\n.d..t...... src/github.com/osism/testbed/netbox/resources/\n<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml\n.d..t...... src/github.com/osism/testbed/network/\n<f..t...... src/github.com/osism/testbed/network/iptables.sh\n<f..t...... src/github.com/osism/testbed/network/vxlan.sh\n.d..t...... src/github.com/osism/testbed/playbooks/\n<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml\n<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml\n<f..t...... src/github.com/osism/testbed/playbooks/post.yml\n<f..t...... src/github.com/osism/testbed/playbooks/pre.yml\n<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml\n.d..t...... src/github.com/osism/testbed/playbooks/templates/\n<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2\n.d..t...... src/github.com/osism/testbed/playbooks/vars/\n<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml\n<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml\n.d..t...... src/github.com/osism/testbed/scripts/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/check.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh\n<f..t...... src/github.com/osism/testbed/scripts/include.sh\n<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh\n<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh\n.d..t...... src/github.com/osism/testbed/scripts/bootstrap/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh\n.d..t...... src/github.com/osism/testbed/scripts/check/\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh\n.d..t...... src/github.com/osism/testbed/scripts/customisations/\n<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh\n.d..t...... src/github.com/osism/testbed/scripts/deploy/\n<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/scripts/upgrade/\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/terraform/\n<f..t...... src/github.com/osism/testbed/terraform/Makefile\n<f..t...... src/github.com/osism/testbed/terraform/README.md\n<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample\n<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/local.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample\n.d..t...... src/github.com/osism/testbed/terraform/environments/\n<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars\n.d..t...... src/github.com/osism/testbed/terraform/scripts/\n<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/\n*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/environment\n*deleting   src/github.com/osism/testbed/terraform/.terraform/\n*deleting   src/github.com/osism/testbed/terraform/variables.tf\n*deleting   src/github.com/osism/testbed/terraform/provider.tf\n*deleting   src/github.com/osism/testbed/terraform/outputs.tf\n*deleting   src/github.com/osism/testbed/terraform/nova.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron.tf\n*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/manager.tf\n*deleting   src/github.com/osism/testbed/terraform/main.tf\n*deleting   src/github.com/osism/testbed/terraform/local.env\n*deleting   src/github.com/osism/testbed/terraform/inventory.ci\n*deleting   src/github.com/osism/testbed/terraform/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/data.tf\n*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars\n*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/.deploy.ci\n*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci\n",
                            "rc": 0,
                            "stdout_lines": [
                                ".d..t...... src/",
                                ".d..t...... src/github.com/",
                                ".d..t...... src/github.com/osism/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/CHANGELOG.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/0b/1a4cc89d202712cb431174fd6158e7682f3b64",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/1a/e6b58d2ab820fda6fc7e19c31f98d3e0c226d4",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/47/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/a3/7092a7e1e69e3523f5770c7c02f4482413eed5",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/e5/8170b7f562ceaf1fbf5908015364fd9de2110d",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-fast.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-fast.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/CHANGELOG.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-services/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD",
                                "<f.st...... src/github.com/osism/ansible-collection-services/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/56/14ee480009ffbd1fe623254aed3aaea46befc8",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/0e94d45a814c75065d42979425a0407f071c3a",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6f/2f892b7b18b7b44e432e38593a701da8667311",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/d2/71a5b882399ba36bbb36181be297a1fc53845c",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_metalbox.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_network.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2",
                                ".d..t...... src/github.com/osism/terraform-base/",
                                "<f..t...... src/github.com/osism/terraform-base/.yamllint.yml",
                                "<f..t...... src/github.com/osism/terraform-base/.zuul.yaml",
                                "<f..t...... src/github.com/osism/terraform-base/LICENSE",
                                "<f..t...... src/github.com/osism/terraform-base/README.md",
                                ".d..t...... src/github.com/osism/terraform-base/.git/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config.worktree",
                                "<f..t...... src/github.com/osism/terraform-base/.git/description",
                                "<f..t...... src/github.com/osism/terraform-base/.git/index",
                                "<f..t...... src/github.com/osism/terraform-base/.git/packed-refs",
                                ".d..t...... src/github.com/osism/terraform-base/.git/branches/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/hooks/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/terraform-base/.git/info/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/info/exclude",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/00/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/01/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/02/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/05/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/06/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/07/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/08/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/10/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/12/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/15/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/16/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/17/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/20/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/21/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/24/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/26/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/29/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/36/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/37/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/41/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/43/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/46/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/47/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/48/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/49/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/52/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/55/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/56/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/58/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/61/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/63/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/66/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/67/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/68/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/70/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/72/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/76/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/80/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/81/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/82/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/83/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/84/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/85/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/86/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/88/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/90/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/91/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/92/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/95/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/96/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/98/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/af/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/da/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/db/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/de/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/df/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/info/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.github/",
                                "<f..t...... src/github.com/osism/terraform-base/.github/renovate.json",
                                ".d..t...... src/github.com/osism/terraform-base/playbooks/",
                                "<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/environments/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf",
                                ".d..t...... src/github.com/osism/testbed/",
                                "<f..t...... src/github.com/osism/testbed/.ansible-lint",
                                "<f..t...... src/github.com/osism/testbed/.flake8",
                                "<f..t...... src/github.com/osism/testbed/.gitignore",
                                "<f..t...... src/github.com/osism/testbed/.yamllint.yml",
                                "<f..t...... src/github.com/osism/testbed/.zuul.yaml",
                                "<f..t...... src/github.com/osism/testbed/LICENSE",
                                "<f..t...... src/github.com/osism/testbed/Makefile",
                                "<f..t...... src/github.com/osism/testbed/README.md",
                                "<f..t...... src/github.com/osism/testbed/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/gilt.yml",
                                "<f..t...... src/github.com/osism/testbed/requirements.txt",
                                "<f..t...... src/github.com/osism/testbed/requirements.yml",
                                "<f..t...... src/github.com/osism/testbed/tox.ini",
                                ".d..t...... src/github.com/osism/testbed/.git/",
                                "<f..t...... src/github.com/osism/testbed/.git/HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/config",
                                "<f..t...... src/github.com/osism/testbed/.git/config.worktree",
                                "<f..t...... src/github.com/osism/testbed/.git/description",
                                "<f.st...... src/github.com/osism/testbed/.git/index",
                                "<f..t...... src/github.com/osism/testbed/.git/packed-refs",
                                ".d..t...... src/github.com/osism/testbed/.git/branches/",
                                ".d..t...... src/github.com/osism/testbed/.git/hooks/",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/testbed/.git/info/",
                                "<f..t...... src/github.com/osism/testbed/.git/info/exclude",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/00/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/00/30c229a586597810f7c072e6e96cc8fec6abd8",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/01/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/02/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/03/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/04/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/05/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/06/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/06/c76461b7ccc166f01b2d33a4628084bb085d9d",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/07/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/08/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/09/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/10/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/11/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/12/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/13/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/14/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/15/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/16/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/17/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/18/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/19/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/19/c1e98eff29413b62f6534ae2c146bf36f85675",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/20/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/21/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/22/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/23/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/24/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/25/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/26/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/27/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/28/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/29/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/30/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/31/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/32/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/33/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/34/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/35/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/36/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/37/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/37/c6b001771fb59f1ffee2c9dc64557ef410bdf0",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/38/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/39/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/40/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/40/4a99b95238a5949445182b00d9725dc2ecd54b",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/41/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/42/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/43/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/44/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/45/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/46/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/47/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/48/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/49/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/50/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/51/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/52/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/53/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/54/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/55/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/56/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/57/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/58/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/59/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5a/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/5a/0516c9d464c12ecd7c4379ea40b1f1ad5e485d",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/60/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/61/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/64/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/65/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/66/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/66/4d846e525baf80e14effbeb2b02e33d9174a5e",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/67/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/68/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/69/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/70/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/71/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/72/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/73/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/74/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/75/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/76/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/76/a4c480e5f6e4bb27df8bda37f74d6388c4321b",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/77/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/78/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/79/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/80/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/81/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/82/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/83/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/84/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/85/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/86/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/87/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/90/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/91/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/92/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/93/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/94/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/95/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/96/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/97/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/98/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/af/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b0/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/b0/a678c098711df8f3fa53ff0af127883997a1a2",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/be/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/da/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/db/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/de/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/df/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f6/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/f6/7fac5930587796a00257dbd2bf7f6faab966e3",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/info/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/testbed/.git/refs/",
                                ".d..t...... src/github.com/osism/testbed/.github/",
                                "<f..t...... src/github.com/osism/testbed/.github/renovate.json",
                                ".d..t...... src/github.com/osism/testbed/.github/workflows/",
                                "<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml",
                                ".d..t...... src/github.com/osism/testbed/ansible/",
                                "<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/README.md",
                                "<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/files/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/daemons",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf",
                                ".d..t...... src/github.com/osism/testbed/contrib/",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py",
                                "<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql",
                                "<f..t...... src/github.com/osism/testbed/contrib/semver2.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py",
                                ".d..t...... src/github.com/osism/testbed/contrib/ansible/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample",
                                ".d..t...... src/github.com/osism/testbed/contrib/assets/",
                                "<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot",
                                ".d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster",
                                ".d..t...... src/github.com/osism/testbed/contrib/openstack-cli/",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example",
                                ".d..t...... src/github.com/osism/testbed/contrib/ownca/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/",
                                "<f..t...... src/github.com/osism/testbed/environments/.vault_pass",
                                "<f..t...... src/github.com/osism/testbed/environments/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/environments/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph.test/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/templates/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2",
                                ".d..t...... src/github.com/osism/testbed/environments/generic/",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf",
                                "<f+++++++++ src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/policy.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/hosts",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/monitoring/",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/openrc",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst",
                                ".d..t...... src/github.com/osism/testbed/inventory/",
                                "<f..t...... src/github.com/osism/testbed/inventory/10-custom",
                                "<f..t...... src/github.com/osism/testbed/inventory/20-roles",
                                "<f..t...... src/github.com/osism/testbed/inventory/99-overwrite",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/all/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/moduletypes/",
                                "<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/netbox/resources/",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml",
                                ".d..t...... src/github.com/osism/testbed/network/",
                                "<f..t...... src/github.com/osism/testbed/network/iptables.sh",
                                "<f..t...... src/github.com/osism/testbed/network/vxlan.sh",
                                ".d..t...... src/github.com/osism/testbed/playbooks/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/post.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/pre.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml",
                                ".d..t...... src/github.com/osism/testbed/playbooks/templates/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2",
                                ".d..t...... src/github.com/osism/testbed/playbooks/vars/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml",
                                ".d..t...... src/github.com/osism/testbed/scripts/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/include.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/bootstrap/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/check/",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/customisations/",
                                "<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/deploy/",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/upgrade/",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/terraform/",
                                "<f..t...... src/github.com/osism/testbed/terraform/Makefile",
                                "<f..t...... src/github.com/osism/testbed/terraform/README.md",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/local.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample",
                                ".d..t...... src/github.com/osism/testbed/terraform/environments/",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars",
                                ".d..t...... src/github.com/osism/testbed/terraform/scripts/",
                                "<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/environment",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/",
                                "*deleting   src/github.com/osism/testbed/terraform/variables.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/provider.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/outputs.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nova.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/main.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/local.env",
                                "*deleting   src/github.com/osism/testbed/terraform/inventory.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/data.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars",
                                "*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.deploy.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-43e6-029b-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:29.194627Z",
                            "start": "2026-03-26T01:37:10.042247Z"
                        },
                        "id": "0242ac1f-6606-43e6-029b-00000000002d",
                        "name": "Synchronize src repos to workspace directory."
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 9,
            "failures": 0,
            "ignored": 0,
            "ok": 26,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/osism/testbed/playbooks/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T01:37:37.069412Z",
                    "start": "2026-03-26T01:37:30.140570Z"
                },
                "id": "0242ac1f-6606-1cd6-a429-000000000002",
                "name": "Pre play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:30.210647Z",
                            "start": "2026-03-26T01:37:30.157761Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:30.254993Z",
                            "start": "2026-03-26T01:37:30.219392Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.484330",
                            "end": "2026-03-26 01:37:32.244534",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1cd6-a429-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:37:30.760204",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-26 01:37:31 - clean up servers\n2026-03-26 01:37:31 - clean up keypairs\n2026-03-26 01:37:31 - wait for servers to be gone\n2026-03-26 01:37:31 - clean up ports\n2026-03-26 01:37:31 - clean up volumes\n2026-03-26 01:37:31 - disconnect routers\n2026-03-26 01:37:31 - clean up subnets\n2026-03-26 01:37:31 - clean up networks\n2026-03-26 01:37:32 - clean up security groups\n2026-03-26 01:37:32 - clean up floating ips\n2026-03-26 01:37:32 - clean up routers",
                            "stdout_lines": [
                                "2026-03-26 01:37:31 - clean up servers",
                                "2026-03-26 01:37:31 - clean up keypairs",
                                "2026-03-26 01:37:31 - wait for servers to be gone",
                                "2026-03-26 01:37:31 - clean up ports",
                                "2026-03-26 01:37:31 - clean up volumes",
                                "2026-03-26 01:37:31 - disconnect routers",
                                "2026-03-26 01:37:31 - clean up subnets",
                                "2026-03-26 01:37:31 - clean up networks",
                                "2026-03-26 01:37:32 - clean up security groups",
                                "2026-03-26 01:37:32 - clean up floating ips",
                                "2026-03-26 01:37:32 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1cd6-a429-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:32.315275Z",
                            "start": "2026-03-26T01:37:30.300850Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000006",
                        "name": "Clean the cloud environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.unarchive",
                            "changed": false,
                            "dest": "/home/zuul-testbed03",
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "handler": "ZipArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": null,
                                    "decrypt": true,
                                    "dest": "/home/zuul-testbed03",
                                    "exclude": [],
                                    "extra_opts": [],
                                    "group": null,
                                    "include": [
                                        "tofu"
                                    ],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/opentofu/opentofu/releases/download/v1.11.5/tofu_1.11.5_linux_amd64.zip",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed03",
                            "size": 4096,
                            "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1774489052.3879364-10-94157175351460/tofu_1.11.5_linux_amd64nkcu2sbv.zip",
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:35.601331Z",
                            "start": "2026-03-26T01:37:32.330756Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000007",
                        "name": "Extract tofu binary"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "rsync",
                                "-avz",
                                "/home/zuul-testbed03/src/github.com/osism/terraform-base/testbed-default/",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform"
                            ],
                            "delta": "0:00:00.061999",
                            "end": "2026-03-26 01:37:35.985963",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "rsync -avz /home/zuul-testbed03/src/github.com/osism/terraform-base/testbed-default/ /home/zuul-testbed03/src/github.com/osism/testbed/terraform",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1cd6-a429-000000000008-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:37:35.923964",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sending incremental file list\ndata.tf\nmain.tf\nmanager.tf\nneutron.tf\nnodes.tf\nnova.tf\noutputs.tf\nprovider.tf\nvariables.tf\ncustomisations/\ncustomisations/access_floatingip_custom.tf\ncustomisations/access_ipv4_custom.tf\ncustomisations/access_ipv6_custom.tf\ncustomisations/default_custom.tf\ncustomisations/external_api_custom.tf\ncustomisations/neutron_floatingip_custom.tf\noverrides/\noverrides/manager_boot_from_image_override.tf\noverrides/manager_boot_from_volume_override.tf\noverrides/neutron_availability_zone_hints_network_override.tf\noverrides/neutron_availability_zone_hints_router_override.tf\noverrides/neutron_router_enable_snat_override.tf\noverrides/nodes_boot_from_image_override.tf\noverrides/nodes_boot_from_volume_override.tf\noverrides/nodes_use_ephemeral_storage_override.tf\n\nsent 6,900 bytes  received 470 bytes  14,740.00 bytes/sec\ntotal size is 26,532  speedup is 3.60",
                            "stdout_lines": [
                                "sending incremental file list",
                                "data.tf",
                                "main.tf",
                                "manager.tf",
                                "neutron.tf",
                                "nodes.tf",
                                "nova.tf",
                                "outputs.tf",
                                "provider.tf",
                                "variables.tf",
                                "customisations/",
                                "customisations/access_floatingip_custom.tf",
                                "customisations/access_ipv4_custom.tf",
                                "customisations/access_ipv6_custom.tf",
                                "customisations/default_custom.tf",
                                "customisations/external_api_custom.tf",
                                "customisations/neutron_floatingip_custom.tf",
                                "overrides/",
                                "overrides/manager_boot_from_image_override.tf",
                                "overrides/manager_boot_from_volume_override.tf",
                                "overrides/neutron_availability_zone_hints_network_override.tf",
                                "overrides/neutron_availability_zone_hints_router_override.tf",
                                "overrides/neutron_router_enable_snat_override.tf",
                                "overrides/nodes_boot_from_image_override.tf",
                                "overrides/nodes_boot_from_volume_override.tf",
                                "overrides/nodes_use_ephemeral_storage_override.tf",
                                "",
                                "sent 6,900 bytes  received 470 bytes  14,740.00 bytes/sec",
                                "total size is 26,532  speedup is 3.60"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1cd6-a429-000000000008-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:36.150028Z",
                            "start": "2026-03-26T01:37:35.636126Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000008",
                        "name": "Sync terraform blueprint"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.template",
                            "changed": true,
                            "checksum": "c7d87fff32e4ed8365964dfc6d78aaf197c62596",
                            "dest": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/local.env",
                            "diff": [],
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "local.env.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "c7d87fff32e4ed8365964dfc6d78aaf197c62596",
                                    "content": null,
                                    "dest": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/local.env",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": "0644",
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1774489056.2255626-19-145795722276754/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "8159794407094435d0565858fa369f4a",
                            "mode": "0644",
                            "owner": "zuul-testbed03",
                            "size": 42,
                            "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1774489056.2255626-19-145795722276754/source",
                            "state": "file",
                            "uid": 1003
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:37.069412Z",
                            "start": "2026-03-26T01:37:36.167936Z"
                        },
                        "id": "0242ac1f-6606-1cd6-a429-000000000009",
                        "name": "Create local.env file"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 3,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/osism/testbed/playbooks/upgrade-stable.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T04:14:59.107669Z",
                    "start": "2026-03-26T01:37:37.974553Z"
                },
                "id": "0242ac1f-6606-ba6a-269d-000000000003",
                "name": "Deploy testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.069967Z",
                            "start": "2026-03-26T01:37:37.990544Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000005",
                        "name": "Print used ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version": "2025.1",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.149166Z",
                            "start": "2026-03-26T01:37:38.081403Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000006",
                        "name": "Print used openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version": "9.5.0",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.224663Z",
                            "start": "2026-03-26T01:37:38.158571Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000007",
                        "name": "Print used manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "image_username": "ubuntu"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.321920Z",
                            "start": "2026-03-26T01:37:38.234188Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000008",
                        "name": "Set facts (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.355079Z",
                            "start": "2026-03-26T01:37:38.330684Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000009",
                        "name": "Set facts (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_manager_version == \"latest\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:37:38.900493Z",
                            "start": "2026-03-26T01:37:38.391112Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000a",
                        "name": "Create infrastructure (latest)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "make",
                                "ENVIRONMENT=ci-ubuntu-24.04",
                                "CLOUD=ci",
                                "IMAGE_USERNAME=ubuntu",
                                "TERRAFORM=/home/zuul-testbed03/tofu",
                                "VERSION_MANAGER=9.5.0",
                                "TEMPEST=false",
                                "CEPH_STACK=ceph-ansible",
                                "create"
                            ],
                            "delta": "0:01:07.805898",
                            "end": "2026-03-26 01:38:47.206862",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "make\n  ENVIRONMENT=ci-ubuntu-24.04\n  CLOUD=ci\n  IMAGE_USERNAME=ubuntu\n  TERRAFORM=/home/zuul-testbed03/tofu\n  VERSION_MANAGER=9.5.0\n  TEMPEST=false\n  CEPH_STACK=ceph-ansible\n  create\n",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000000b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:37:39.400964",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\nInitializing the backend...\n\nInitializing provider plugins...\n- terraform.io/builtin/terraform is built in to OpenTofu\n- Finding hashicorp/local versions matching \">= 2.2.0\"...\n- Finding latest version of hashicorp/null...\n- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...\n- Installing hashicorp/local v2.7.0...\n- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)\n- Installing hashicorp/null v3.2.4...\n- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)\n- Installing terraform-provider-openstack/openstack v3.4.0...\n- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)\n\nProviders are signed by their developers.\nIf you'd like to know more about provider signing, you can read about it here:\nhttps://opentofu.org/docs/cli/plugins/signing/\n\nOpenTofu has created a lock file .terraform.lock.hcl to record the provider\nselections it made above. Include this file in your version control repository\nso that OpenTofu can guarantee to make the same selections by default when\nyou run \"tofu init\" in the future.\n\nOpenTofu has been successfully initialized!\n\nYou may now begin working with OpenTofu. Try running \"tofu plan\" to see\nany changes that are required for your infrastructure. All OpenTofu commands\nshould now work.\n\nIf you ever set or change modules or backend configuration for OpenTofu,\nrerun this command to reinitialize your working directory. If you forget, other\ncommands will detect it and remind you to do so if necessary.\nCreated and switched to workspace \"ci\"!\n\nYou're now on a new, empty workspace. Workspaces isolate their state,\nso if you run \"tofu plan\" OpenTofu will not see any existing state\nfor this configuration.\nci.auto.tfvars\ndefault_custom.tf\ndata.openstack_networking_network_v2.public: Reading...\ndata.openstack_networking_network_v2.public: Read complete after 0s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]\n\nOpenTofu used the selected providers to generate the following execution\nplan. Resource actions are indicated with the following symbols:\n  + create\n <= read (data resources)\n\nOpenTofu will perform the following actions:\n\n  # data.openstack_images_image_v2.image will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # data.openstack_images_image_v2.image_node will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image_node\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # local_file.MANAGER_ADDRESS will be created\n  + resource \"local_file\" \"MANAGER_ADDRESS\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".MANAGER_ADDRESS.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.id_rsa_pub will be created\n  + resource \"local_file\" \"id_rsa_pub\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".id_rsa.ci.pub\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.inventory will be created\n  + resource \"local_file\" \"inventory\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \"inventory.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_sensitive_file.id_rsa will be created\n  + resource \"local_sensitive_file\" \"id_rsa\" {\n      + content              = (sensitive value)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0700\"\n      + file_permission      = \"0600\"\n      + filename             = \".id_rsa.ci\"\n      + id                   = (known after apply)\n    }\n\n  # null_resource.node_semaphore will be created\n  + resource \"null_resource\" \"node_semaphore\" {\n      + id = (known after apply)\n    }\n\n  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-manager-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[6] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-6-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[7] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-7-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[8] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-8-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_compute_instance_v2.manager_server will be created\n  + resource \"openstack_compute_instance_v2\" \"manager_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-4V-16\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-manager\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = (sensitive value)\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[0] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-0\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[1] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-1\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[2] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-2\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[3] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-3\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[4] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-4\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[5] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-5\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_keypair_v2.key will be created\n  + resource \"openstack_compute_keypair_v2\" \"key\" {\n      + fingerprint = (known after apply)\n      + id          = (known after apply)\n      + name        = \"testbed\"\n      + private_key = (sensitive value)\n      + public_key  = (known after apply)\n      + region      = (known after apply)\n      + user_id     = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created\n  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {\n      + fixed_ip    = (known after apply)\n      + floating_ip = (known after apply)\n      + id          = (known after apply)\n      + port_id     = (known after apply)\n      + region      = (known after apply)\n    }\n\n  # openstack_networking_floatingip_v2.manager_floating_ip will be created\n  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {\n      + address    = (known after apply)\n      + all_tags   = (known after apply)\n      + dns_domain = (known after apply)\n      + dns_name   = (known after apply)\n      + fixed_ip   = (known after apply)\n      + id         = (known after apply)\n      + pool       = \"public\"\n      + port_id    = (known after apply)\n      + region     = (known after apply)\n      + subnet_id  = (known after apply)\n      + tenant_id  = (known after apply)\n    }\n\n  # openstack_networking_network_v2.net_management will be created\n  + resource \"openstack_networking_network_v2\" \"net_management\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + dns_domain              = (known after apply)\n      + external                = (known after apply)\n      + id                      = (known after apply)\n      + mtu                     = (known after apply)\n      + name                    = \"net-testbed-management\"\n      + port_security_enabled   = (known after apply)\n      + qos_policy_id           = (known after apply)\n      + region                  = (known after apply)\n      + shared                  = (known after apply)\n      + tenant_id               = (known after apply)\n      + transparent_vlan        = (known after apply)\n\n      + segments (known after apply)\n    }\n\n  # openstack_networking_port_v2.manager_port_management will be created\n  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.5\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[0] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.10\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[1] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.11\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[2] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.12\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[3] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.13\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[4] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.14\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[5] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.15\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_router_interface_v2.router_interface will be created\n  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {\n      + force_destroy = false\n      + id            = (known after apply)\n      + port_id       = (known after apply)\n      + region        = (known after apply)\n      + router_id     = (known after apply)\n      + subnet_id     = (known after apply)\n    }\n\n  # openstack_networking_router_v2.router will be created\n  + resource \"openstack_networking_router_v2\" \"router\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + distributed             = (known after apply)\n      + enable_snat             = (known after apply)\n      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"\n      + external_qos_policy_id  = (known after apply)\n      + id                      = (known after apply)\n      + name                    = \"testbed\"\n      + region                  = (known after apply)\n      + tenant_id               = (known after apply)\n\n      + external_fixed_ip (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {\n      + description             = \"ssh\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 22\n      + port_range_min          = 22\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {\n      + description             = \"wireguard\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 51820\n      + port_range_min          = 51820\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {\n      + description             = \"vrrp\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"112\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_management will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {\n      + all_tags    = (known after apply)\n      + description = \"management security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-management\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_node will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {\n      + all_tags    = (known after apply)\n      + description = \"node security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-node\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_subnet_v2.subnet_management will be created\n  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {\n      + all_tags          = (known after apply)\n      + cidr              = \"192.168.16.0/20\"\n      + dns_nameservers   = [\n          + \"8.8.8.8\",\n          + \"9.9.9.9\",\n        ]\n      + enable_dhcp       = true\n      + gateway_ip        = (known after apply)\n      + id                = (known after apply)\n      + ip_version        = 4\n      + ipv6_address_mode = (known after apply)\n      + ipv6_ra_mode      = (known after apply)\n      + name              = \"subnet-testbed-management\"\n      + network_id        = (known after apply)\n      + no_gateway        = false\n      + region            = (known after apply)\n      + service_types     = (known after apply)\n      + tenant_id         = (known after apply)\n\n      + allocation_pool {\n          + end   = \"192.168.31.250\"\n          + start = \"192.168.31.200\"\n        }\n    }\n\n  # terraform_data.image will be created\n  + resource \"terraform_data\" \"image\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\n  # terraform_data.image_node will be created\n  + resource \"terraform_data\" \"image_node\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\nPlan: 64 to add, 0 to change, 0 to destroy.\n\nChanges to Outputs:\n  + manager_address = (sensitive value)\n  + private_key     = (sensitive value)\nterraform_data.image: Creating...\nterraform_data.image_node: Creating...\nterraform_data.image_node: Creation complete after 0s [id=08fb72f6-2178-02dd-a5f1-d4706cb277dc]\nterraform_data.image: Creation complete after 0s [id=8d4cdb9a-566a-9f81-21f9-f3b0207d51e9]\ndata.openstack_images_image_v2.image_node: Reading...\nopenstack_blockstorage_volume_v3.node_volume[6]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creating...\nopenstack_compute_keypair_v2.key: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creating...\nopenstack_networking_network_v2.net_management: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creating...\ndata.openstack_images_image_v2.image_node: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creating...\nopenstack_compute_keypair_v2.key: Creation complete after 0s [id=testbed]\nopenstack_blockstorage_volume_v3.node_volume[8]: Creating...\nopenstack_networking_network_v2.net_management: Creation complete after 1s [id=7977b382-565f-4526-9bfa-099755a02bf8]\ndata.openstack_images_image_v2.image: Reading...\ndata.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 3s [id=d11e4e4a-db1d-44df-8da9-5de7e993dd80]\nopenstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 3s [id=47760649-09e9-4ed8-8303-e5ee473a8102]\nopenstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 3s [id=7e352b46-e023-45cf-8a88-51cc46240a44]\nlocal_sensitive_file.id_rsa: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creating...\nlocal_sensitive_file.id_rsa: Creation complete after 0s [id=ff9bca06fb0c8ea2aa09bf77e2c39317213b84ec]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=2dae49df-17cb-48b5-9940-ec5e7ec792d8]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 3s [id=863ba5d2-7e2f-4393-95a6-83543745d331]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 3s [id=943c088c-5b56-4173-ab64-ec81e1cc816d]\nopenstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=a52ec37c-b4ea-4f83-9b16-3c0f6ce85263]\nopenstack_networking_subnet_v2.subnet_management: Creating...\nlocal_file.id_rsa_pub: Creating...\nlocal_file.id_rsa_pub: Creation complete after 0s [id=66f2e27fe84d66c13d33c369431a63d3178daf01]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=7db5f133-fe7b-42a4-ad57-b076dc1856ab]\nopenstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 3s [id=8ddd7966-84e6-4951-8a08-7b4fb4af2bd2]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=c374eb4c-3572-4f0b-927c-38d35765f44a]\nopenstack_networking_subnet_v2.subnet_management: Creation complete after 2s [id=e73ccf28-9912-4c70-ab14-4cc9f269ba17]\nopenstack_networking_router_v2.router: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 4s [id=7634648a-b5a4-45bc-ac0b-8484a2642b22]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 4s [id=48d73a84-835d-480a-92c3-3edf7ed142ea]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 4s [id=ce600cf2-62c4-44aa-8248-5535335c6519]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 4s [id=4fa924fa-33d9-43ce-b208-159d6f6ab539]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 4s [id=2e41bcf9-ad92-42bb-b49e-289ca95def9f]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=2ece1d7d-b762-44e0-80cf-0ec8d4e65a06]\nopenstack_networking_router_v2.router: Creation complete after 3s [id=458147c9-9b72-430f-8b8f-65d092fea4f7]\nopenstack_networking_router_interface_v2.router_interface: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creation complete after 0s [id=9a5a0685-9793-4191-a4d7-ff62567ebcb4]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=cb38da93-75ae-4123-a24c-4683e394bfd4]\nopenstack_networking_port_v2.manager_port_management: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 0s [id=c2a1604f-4c2f-48df-9b6a-44bdc3f502d4]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=44bd8a20-d120-46b1-8e0b-a5cf643511cf]\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 0s [id=e4189816-f698-4e66-9031-d3b5023525e2]\nopenstack_networking_port_v2.node_port_management[5]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 0s [id=eecaeda2-c988-4496-836a-5e4bbf345ad6]\nopenstack_networking_port_v2.node_port_management[4]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=a7d71b60-6ba9-4719-80a0-0e386a7ff561]\nopenstack_networking_port_v2.node_port_management[3]: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=d0917be5-d774-4ae8-aa8e-88d5b51d1da3]\nopenstack_networking_port_v2.node_port_management[0]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=c25c7dd1-b378-4088-95fa-0599b351c201]\nopenstack_networking_port_v2.node_port_management[2]: Creating...\nopenstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=5d571e2f-b701-4263-883f-e7221cc0c90a]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=2f203c7d-ac9e-403c-a339-ce95ce237fc4]\nopenstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=0d011944-d664-48dc-afd5-e876547c06cd]\nopenstack_networking_port_v2.node_port_management[0]: Creation complete after 0s [id=4acb0156-c5c7-4457-a252-e47665f1efbd]\nopenstack_networking_port_v2.node_port_management[3]: Creation complete after 0s [id=da761b17-04fe-4927-b0ae-964e0d48a3d8]\nopenstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=be5d018b-5284-4a13-84e4-e85247bdd4a2]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=41231804-f8cb-4307-a0c3-4120acada0c0]\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 1s [id=495645ce-cf06-4d1e-8843-90be4513fa23]\nopenstack_networking_port_v2.node_port_management[2]: Creation complete after 0s [id=c09a035e-b37c-4623-b251-1a605583e2e5]\nopenstack_networking_router_interface_v2.router_interface: Creation complete after 2s [id=ac1dfea1-1e14-4018-8481-5e4ce13cebb6]\nopenstack_networking_floatingip_v2.manager_floating_ip: Creating...\nopenstack_compute_instance_v2.node_server[5]: Creating...\nopenstack_compute_instance_v2.node_server[4]: Creating...\nopenstack_compute_instance_v2.node_server[0]: Creating...\nopenstack_compute_instance_v2.node_server[2]: Creating...\nopenstack_compute_instance_v2.node_server[1]: Creating...\nopenstack_compute_instance_v2.node_server[3]: Creating...\nopenstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=abd7a441-ee41-4e76-ab04-3935b7055fc6]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...\nlocal_file.MANAGER_ADDRESS: Creating...\nlocal_file.MANAGER_ADDRESS: Creation complete after 0s [id=ec9eca104d958caee09087975d6798eb0bb86b60]\nlocal_file.inventory: Creating...\nlocal_file.inventory: Creation complete after 0s [id=30d1748c77c82795d6e88e4b1ba40d79e0a8dfb3]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 1s [id=abd7a441-ee41-4e76-ab04-3935b7055fc6]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Creation complete after 21s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816]\nopenstack_compute_instance_v2.node_server[2]: Creation complete after 21s [id=df4dc442-7268-46a7-a597-7c64799a6d9c]\nopenstack_compute_instance_v2.node_server[0]: Creation complete after 21s [id=70a77876-4543-4e2a-b0a0-13586b389da7]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Creation complete after 31s [id=827668dc-3e5a-4728-aa0b-89be81db88d7]\nopenstack_compute_instance_v2.node_server[1]: Creation complete after 32s [id=8b5c9d75-54b2-4ca9-914e-7a0667bc48a0]\nopenstack_compute_instance_v2.node_server[5]: Creation complete after 32s [id=65bc15c7-423f-40be-804d-198c847c7c7d]\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...\nnull_resource.node_semaphore: Creating...\nnull_resource.node_semaphore: Creation complete after 0s [id=268805636366249020]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...\nopenstack_compute_instance_v2.manager_server: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 3s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/7e352b46-e023-45cf-8a88-51cc46240a44]\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 3s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/2dae49df-17cb-48b5-9940-ec5e7ec792d8]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 3s [id=65bc15c7-423f-40be-804d-198c847c7c7d/8ddd7966-84e6-4951-8a08-7b4fb4af2bd2]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 3s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/a52ec37c-b4ea-4f83-9b16-3c0f6ce85263]\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 3s [id=65bc15c7-423f-40be-804d-198c847c7c7d/47760649-09e9-4ed8-8303-e5ee473a8102]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 3s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/863ba5d2-7e2f-4393-95a6-83543745d331]\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 9s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/7db5f133-fe7b-42a4-ad57-b076dc1856ab]\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 9s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/d11e4e4a-db1d-44df-8da9-5de7e993dd80]\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 9s [id=65bc15c7-423f-40be-804d-198c847c7c7d/943c088c-5b56-4173-ab64-ec81e1cc816d]\nopenstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.manager_server: Creation complete after 20s [id=58a35edb-618b-4cc9-abf2-21320a733826]\n\nApply complete! Resources: 64 added, 0 changed, 0 destroyed.\n\nOutputs:\n\nmanager_address = <sensitive>\nprivate_key = <sensitive>",
                            "stdout_lines": [
                                "",
                                "Initializing the backend...",
                                "",
                                "Initializing provider plugins...",
                                "- terraform.io/builtin/terraform is built in to OpenTofu",
                                "- Finding hashicorp/local versions matching \">= 2.2.0\"...",
                                "- Finding latest version of hashicorp/null...",
                                "- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...",
                                "- Installing hashicorp/local v2.7.0...",
                                "- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing hashicorp/null v3.2.4...",
                                "- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing terraform-provider-openstack/openstack v3.4.0...",
                                "- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)",
                                "",
                                "Providers are signed by their developers.",
                                "If you'd like to know more about provider signing, you can read about it here:",
                                "https://opentofu.org/docs/cli/plugins/signing/",
                                "",
                                "OpenTofu has created a lock file .terraform.lock.hcl to record the provider",
                                "selections it made above. Include this file in your version control repository",
                                "so that OpenTofu can guarantee to make the same selections by default when",
                                "you run \"tofu init\" in the future.",
                                "",
                                "OpenTofu has been successfully initialized!",
                                "",
                                "You may now begin working with OpenTofu. Try running \"tofu plan\" to see",
                                "any changes that are required for your infrastructure. All OpenTofu commands",
                                "should now work.",
                                "",
                                "If you ever set or change modules or backend configuration for OpenTofu,",
                                "rerun this command to reinitialize your working directory. If you forget, other",
                                "commands will detect it and remind you to do so if necessary.",
                                "Created and switched to workspace \"ci\"!",
                                "",
                                "You're now on a new, empty workspace. Workspaces isolate their state,",
                                "so if you run \"tofu plan\" OpenTofu will not see any existing state",
                                "for this configuration.",
                                "ci.auto.tfvars",
                                "default_custom.tf",
                                "data.openstack_networking_network_v2.public: Reading...",
                                "data.openstack_networking_network_v2.public: Read complete after 0s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]",
                                "",
                                "OpenTofu used the selected providers to generate the following execution",
                                "plan. Resource actions are indicated with the following symbols:",
                                "  + create",
                                " <= read (data resources)",
                                "",
                                "OpenTofu will perform the following actions:",
                                "",
                                "  # data.openstack_images_image_v2.image will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # data.openstack_images_image_v2.image_node will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image_node\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.MANAGER_ADDRESS will be created",
                                "  + resource \"local_file\" \"MANAGER_ADDRESS\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".MANAGER_ADDRESS.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.id_rsa_pub will be created",
                                "  + resource \"local_file\" \"id_rsa_pub\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".id_rsa.ci.pub\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.inventory will be created",
                                "  + resource \"local_file\" \"inventory\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \"inventory.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_sensitive_file.id_rsa will be created",
                                "  + resource \"local_sensitive_file\" \"id_rsa\" {",
                                "      + content              = (sensitive value)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0700\"",
                                "      + file_permission      = \"0600\"",
                                "      + filename             = \".id_rsa.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # null_resource.node_semaphore will be created",
                                "  + resource \"null_resource\" \"node_semaphore\" {",
                                "      + id = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-manager-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[6] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-6-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[7] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-7-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[8] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-8-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.manager_server will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"manager_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-4V-16\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-manager\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = (sensitive value)",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[0] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-0\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[1] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-1\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[2] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-2\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[3] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-3\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[4] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-4\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[5] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-5\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_keypair_v2.key will be created",
                                "  + resource \"openstack_compute_keypair_v2\" \"key\" {",
                                "      + fingerprint = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed\"",
                                "      + private_key = (sensitive value)",
                                "      + public_key  = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + user_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created",
                                "  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {",
                                "      + fixed_ip    = (known after apply)",
                                "      + floating_ip = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + port_id     = (known after apply)",
                                "      + region      = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_v2.manager_floating_ip will be created",
                                "  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {",
                                "      + address    = (known after apply)",
                                "      + all_tags   = (known after apply)",
                                "      + dns_domain = (known after apply)",
                                "      + dns_name   = (known after apply)",
                                "      + fixed_ip   = (known after apply)",
                                "      + id         = (known after apply)",
                                "      + pool       = \"public\"",
                                "      + port_id    = (known after apply)",
                                "      + region     = (known after apply)",
                                "      + subnet_id  = (known after apply)",
                                "      + tenant_id  = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_network_v2.net_management will be created",
                                "  + resource \"openstack_networking_network_v2\" \"net_management\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + dns_domain              = (known after apply)",
                                "      + external                = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + mtu                     = (known after apply)",
                                "      + name                    = \"net-testbed-management\"",
                                "      + port_security_enabled   = (known after apply)",
                                "      + qos_policy_id           = (known after apply)",
                                "      + region                  = (known after apply)",
                                "      + shared                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "      + transparent_vlan        = (known after apply)",
                                "",
                                "      + segments (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.manager_port_management will be created",
                                "  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.5\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[0] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.10\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[1] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.11\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[2] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.12\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[3] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.13\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[4] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.14\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[5] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.15\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_router_interface_v2.router_interface will be created",
                                "  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {",
                                "      + force_destroy = false",
                                "      + id            = (known after apply)",
                                "      + port_id       = (known after apply)",
                                "      + region        = (known after apply)",
                                "      + router_id     = (known after apply)",
                                "      + subnet_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_router_v2.router will be created",
                                "  + resource \"openstack_networking_router_v2\" \"router\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + distributed             = (known after apply)",
                                "      + enable_snat             = (known after apply)",
                                "      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"",
                                "      + external_qos_policy_id  = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + name                    = \"testbed\"",
                                "      + region                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "",
                                "      + external_fixed_ip (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {",
                                "      + description             = \"ssh\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 22",
                                "      + port_range_min          = 22",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {",
                                "      + description             = \"wireguard\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 51820",
                                "      + port_range_min          = 51820",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {",
                                "      + description             = \"vrrp\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"112\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_management will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"management security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-management\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_node will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"node security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-node\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_subnet_v2.subnet_management will be created",
                                "  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {",
                                "      + all_tags          = (known after apply)",
                                "      + cidr              = \"192.168.16.0/20\"",
                                "      + dns_nameservers   = [",
                                "          + \"8.8.8.8\",",
                                "          + \"9.9.9.9\",",
                                "        ]",
                                "      + enable_dhcp       = true",
                                "      + gateway_ip        = (known after apply)",
                                "      + id                = (known after apply)",
                                "      + ip_version        = 4",
                                "      + ipv6_address_mode = (known after apply)",
                                "      + ipv6_ra_mode      = (known after apply)",
                                "      + name              = \"subnet-testbed-management\"",
                                "      + network_id        = (known after apply)",
                                "      + no_gateway        = false",
                                "      + region            = (known after apply)",
                                "      + service_types     = (known after apply)",
                                "      + tenant_id         = (known after apply)",
                                "",
                                "      + allocation_pool {",
                                "          + end   = \"192.168.31.250\"",
                                "          + start = \"192.168.31.200\"",
                                "        }",
                                "    }",
                                "",
                                "  # terraform_data.image will be created",
                                "  + resource \"terraform_data\" \"image\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "  # terraform_data.image_node will be created",
                                "  + resource \"terraform_data\" \"image_node\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "Plan: 64 to add, 0 to change, 0 to destroy.",
                                "",
                                "Changes to Outputs:",
                                "  + manager_address = (sensitive value)",
                                "  + private_key     = (sensitive value)",
                                "terraform_data.image: Creating...",
                                "terraform_data.image_node: Creating...",
                                "terraform_data.image_node: Creation complete after 0s [id=08fb72f6-2178-02dd-a5f1-d4706cb277dc]",
                                "terraform_data.image: Creation complete after 0s [id=8d4cdb9a-566a-9f81-21f9-f3b0207d51e9]",
                                "data.openstack_images_image_v2.image_node: Reading...",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creating...",
                                "openstack_compute_keypair_v2.key: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creating...",
                                "openstack_networking_network_v2.net_management: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creating...",
                                "data.openstack_images_image_v2.image_node: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creating...",
                                "openstack_compute_keypair_v2.key: Creation complete after 0s [id=testbed]",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creating...",
                                "openstack_networking_network_v2.net_management: Creation complete after 1s [id=7977b382-565f-4526-9bfa-099755a02bf8]",
                                "data.openstack_images_image_v2.image: Reading...",
                                "data.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 3s [id=d11e4e4a-db1d-44df-8da9-5de7e993dd80]",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 3s [id=47760649-09e9-4ed8-8303-e5ee473a8102]",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 3s [id=7e352b46-e023-45cf-8a88-51cc46240a44]",
                                "local_sensitive_file.id_rsa: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creating...",
                                "local_sensitive_file.id_rsa: Creation complete after 0s [id=ff9bca06fb0c8ea2aa09bf77e2c39317213b84ec]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=2dae49df-17cb-48b5-9940-ec5e7ec792d8]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 3s [id=863ba5d2-7e2f-4393-95a6-83543745d331]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 3s [id=943c088c-5b56-4173-ab64-ec81e1cc816d]",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=a52ec37c-b4ea-4f83-9b16-3c0f6ce85263]",
                                "openstack_networking_subnet_v2.subnet_management: Creating...",
                                "local_file.id_rsa_pub: Creating...",
                                "local_file.id_rsa_pub: Creation complete after 0s [id=66f2e27fe84d66c13d33c369431a63d3178daf01]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=7db5f133-fe7b-42a4-ad57-b076dc1856ab]",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 3s [id=8ddd7966-84e6-4951-8a08-7b4fb4af2bd2]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=c374eb4c-3572-4f0b-927c-38d35765f44a]",
                                "openstack_networking_subnet_v2.subnet_management: Creation complete after 2s [id=e73ccf28-9912-4c70-ab14-4cc9f269ba17]",
                                "openstack_networking_router_v2.router: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 4s [id=7634648a-b5a4-45bc-ac0b-8484a2642b22]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 4s [id=48d73a84-835d-480a-92c3-3edf7ed142ea]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 4s [id=ce600cf2-62c4-44aa-8248-5535335c6519]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 4s [id=4fa924fa-33d9-43ce-b208-159d6f6ab539]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 4s [id=2e41bcf9-ad92-42bb-b49e-289ca95def9f]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=2ece1d7d-b762-44e0-80cf-0ec8d4e65a06]",
                                "openstack_networking_router_v2.router: Creation complete after 3s [id=458147c9-9b72-430f-8b8f-65d092fea4f7]",
                                "openstack_networking_router_interface_v2.router_interface: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creation complete after 0s [id=9a5a0685-9793-4191-a4d7-ff62567ebcb4]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=cb38da93-75ae-4123-a24c-4683e394bfd4]",
                                "openstack_networking_port_v2.manager_port_management: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 0s [id=c2a1604f-4c2f-48df-9b6a-44bdc3f502d4]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=44bd8a20-d120-46b1-8e0b-a5cf643511cf]",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 0s [id=e4189816-f698-4e66-9031-d3b5023525e2]",
                                "openstack_networking_port_v2.node_port_management[5]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 0s [id=eecaeda2-c988-4496-836a-5e4bbf345ad6]",
                                "openstack_networking_port_v2.node_port_management[4]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=a7d71b60-6ba9-4719-80a0-0e386a7ff561]",
                                "openstack_networking_port_v2.node_port_management[3]: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=d0917be5-d774-4ae8-aa8e-88d5b51d1da3]",
                                "openstack_networking_port_v2.node_port_management[0]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=c25c7dd1-b378-4088-95fa-0599b351c201]",
                                "openstack_networking_port_v2.node_port_management[2]: Creating...",
                                "openstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=5d571e2f-b701-4263-883f-e7221cc0c90a]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=2f203c7d-ac9e-403c-a339-ce95ce237fc4]",
                                "openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=0d011944-d664-48dc-afd5-e876547c06cd]",
                                "openstack_networking_port_v2.node_port_management[0]: Creation complete after 0s [id=4acb0156-c5c7-4457-a252-e47665f1efbd]",
                                "openstack_networking_port_v2.node_port_management[3]: Creation complete after 0s [id=da761b17-04fe-4927-b0ae-964e0d48a3d8]",
                                "openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=be5d018b-5284-4a13-84e4-e85247bdd4a2]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=41231804-f8cb-4307-a0c3-4120acada0c0]",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 1s [id=495645ce-cf06-4d1e-8843-90be4513fa23]",
                                "openstack_networking_port_v2.node_port_management[2]: Creation complete after 0s [id=c09a035e-b37c-4623-b251-1a605583e2e5]",
                                "openstack_networking_router_interface_v2.router_interface: Creation complete after 2s [id=ac1dfea1-1e14-4018-8481-5e4ce13cebb6]",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creating...",
                                "openstack_compute_instance_v2.node_server[5]: Creating...",
                                "openstack_compute_instance_v2.node_server[4]: Creating...",
                                "openstack_compute_instance_v2.node_server[0]: Creating...",
                                "openstack_compute_instance_v2.node_server[2]: Creating...",
                                "openstack_compute_instance_v2.node_server[1]: Creating...",
                                "openstack_compute_instance_v2.node_server[3]: Creating...",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=abd7a441-ee41-4e76-ab04-3935b7055fc6]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...",
                                "local_file.MANAGER_ADDRESS: Creating...",
                                "local_file.MANAGER_ADDRESS: Creation complete after 0s [id=ec9eca104d958caee09087975d6798eb0bb86b60]",
                                "local_file.inventory: Creating...",
                                "local_file.inventory: Creation complete after 0s [id=30d1748c77c82795d6e88e4b1ba40d79e0a8dfb3]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 1s [id=abd7a441-ee41-4e76-ab04-3935b7055fc6]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Creation complete after 21s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816]",
                                "openstack_compute_instance_v2.node_server[2]: Creation complete after 21s [id=df4dc442-7268-46a7-a597-7c64799a6d9c]",
                                "openstack_compute_instance_v2.node_server[0]: Creation complete after 21s [id=70a77876-4543-4e2a-b0a0-13586b389da7]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Creation complete after 31s [id=827668dc-3e5a-4728-aa0b-89be81db88d7]",
                                "openstack_compute_instance_v2.node_server[1]: Creation complete after 32s [id=8b5c9d75-54b2-4ca9-914e-7a0667bc48a0]",
                                "openstack_compute_instance_v2.node_server[5]: Creation complete after 32s [id=65bc15c7-423f-40be-804d-198c847c7c7d]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...",
                                "null_resource.node_semaphore: Creating...",
                                "null_resource.node_semaphore: Creation complete after 0s [id=268805636366249020]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...",
                                "openstack_compute_instance_v2.manager_server: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 3s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/7e352b46-e023-45cf-8a88-51cc46240a44]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 3s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/2dae49df-17cb-48b5-9940-ec5e7ec792d8]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 3s [id=65bc15c7-423f-40be-804d-198c847c7c7d/8ddd7966-84e6-4951-8a08-7b4fb4af2bd2]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 3s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/a52ec37c-b4ea-4f83-9b16-3c0f6ce85263]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 3s [id=65bc15c7-423f-40be-804d-198c847c7c7d/47760649-09e9-4ed8-8303-e5ee473a8102]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 3s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/863ba5d2-7e2f-4393-95a6-83543745d331]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 9s [id=471ae1e9-73d9-4b08-8a76-ea77edb20816/7db5f133-fe7b-42a4-ad57-b076dc1856ab]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 9s [id=827668dc-3e5a-4728-aa0b-89be81db88d7/d11e4e4a-db1d-44df-8da9-5de7e993dd80]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 9s [id=65bc15c7-423f-40be-804d-198c847c7c7d/943c088c-5b56-4173-ab64-ec81e1cc816d]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Creation complete after 20s [id=58a35edb-618b-4cc9-abf2-21320a733826]",
                                "",
                                "Apply complete! Resources: 64 added, 0 changed, 0 destroyed.",
                                "",
                                "Outputs:",
                                "",
                                "manager_address = <sensitive>",
                                "private_key = <sensitive>"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000000b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:38:47.520781Z",
                            "start": "2026-03-26T01:37:38.941063Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000b",
                        "name": "Create infrastructure (stable)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTIuNTQK",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:38:48.044490Z",
                            "start": "2026-03-26T01:38:47.536365Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000c",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.192.54"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:38:48.137947Z",
                            "start": "2026-03-26T01:38:48.055212Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000d",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed03/src/github.com/osism/ansible-collection-commons"
                                    ],
                                    "delta": "0:00:00.648358",
                                    "end": "2026-03-26 01:38:49.084437",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed03/src/github.com/osism/ansible-collection-commons\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed03",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "commons",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-26 01:38:48.436079",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n[WARNING]: Collection osism.services does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.commons:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons'\nCreated collection for osism.commons:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons\nosism.commons:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.commons:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons'",
                                        "Created collection for osism.commons:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons",
                                        "osism.commons:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed03/src/github.com/osism/ansible-collection-services"
                                    ],
                                    "delta": "0:00:00.676251",
                                    "end": "2026-03-26 01:38:50.066163",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed03/src/github.com/osism/ansible-collection-services\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed03",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "services",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-26 01:38:49.389912",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.services does not support Ansible version 2.15.2\n[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.services:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services'",
                                        "Created collection for osism.services:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services",
                                        "osism.services:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:38:50.129251Z",
                            "start": "2026-03-26T01:38:48.146642Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000e",
                        "name": "Update ansible collections"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 10,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.54",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:39:00.752093Z",
                            "start": "2026-03-26T01:38:50.144469Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000000f",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-26 01:39:00.786743",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-26 01:40:00.789808",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:40:00.809349Z",
                            "start": "2026-03-26T01:39:00.762290Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000010",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:40:02.401632Z",
                            "start": "2026-03-26T01:40:00.884820Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000011",
                        "name": "Fetch manager ssh hostkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "cp /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id* .",
                            "delta": "0:00:00.010601",
                            "end": "2026-03-26 01:40:02.772101",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "cp /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id* .",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000012-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:40:02.761500",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000012-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:40:02.968653Z",
                            "start": "2026-03-26T01:40:02.455654Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000012",
                        "name": "Get ssh keypair from terraform environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:40:03.024524Z",
                            "start": "2026-03-26T01:40:02.983310Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000013",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                            "delta": "0:06:14.895732",
                            "end": "2026-03-26 01:46:18.239539",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000014-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:40:03.343807",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Wait for cloud-init to finish] *******************************************\n\nTASK [Check /var/lib/cloud/instance/boot-finished] *****************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Update package cache] ****************************************************\nskipping: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nskipping: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Fail if Ubuntu version is lower than 24.04] ******************************\nskipping: [testbed-manager]\n\nTASK [Fail if Debian version is lower than 12] *********************************\nskipping: [testbed-manager]\n\nTASK [Set APT options on manager] **********************************************\nchanged: [testbed-manager]\n\nTASK [Update APT cache and run dist-upgrade] ***********************************\nchanged: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nchanged: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nchanged: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [Get current user] ********************************************************\nok: [testbed-manager]\n\nTASK [Create venv directory] ***************************************************\nchanged: [testbed-manager]\n\nTASK [Install netaddr in venv] *************************************************\nchanged: [testbed-manager]\n\nTASK [Install ansible-core in venv] ********************************************\nchanged: [testbed-manager]\n\nTASK [Install requests >= 2.32.2] **********************************************\nchanged: [testbed-manager]\n\nTASK [Install docker >= 7.1.0] *************************************************\nchanged: [testbed-manager]\n\nTASK [Create directories in /opt/src] ******************************************\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Sync sources in /opt/src] ************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Create /usr/share/ansible directory] *************************************\nchanged: [testbed-manager]\n\nTASK [Install collections from Ansible galaxy] *********************************\nchanged: [testbed-manager] => (item=ansible.netcommon)\nchanged: [testbed-manager] => (item=ansible.posix)\nchanged: [testbed-manager] => (item=community.docker>=3.10.2)\n\nTASK [Install local collections] ***********************************************\nchanged: [testbed-manager] => (item=ansible-collection-commons)\nchanged: [testbed-manager] => (item=ansible-collection-services)\n\nPLAY [Create operator user] ****************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Create operator group] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Create user] ************************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nchanged: [testbed-manager] => (item=adm)\nchanged: [testbed-manager] => (item=sudo)\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nchanged: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)\nchanged: [testbed-manager] => (item=export LANG=C.UTF-8)\nchanged: [testbed-manager] => (item=export LC_ALL=C.UTF-8)\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set password] ***********************************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Recursively change ownership of /opt/venv] *******************************\nchanged: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Wait for cloud-init to finish] *******************************************",
                                "",
                                "TASK [Check /var/lib/cloud/instance/boot-finished] *****************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Update package cache] ****************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Ubuntu version is lower than 24.04] ******************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Debian version is lower than 12] *********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set APT options on manager] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Update APT cache and run dist-upgrade] ***********************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get current user] ********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Create venv directory] ***************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install netaddr in venv] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install ansible-core in venv] ********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install requests >= 2.32.2] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install docker >= 7.1.0] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create directories in /opt/src] ******************************************",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Sync sources in /opt/src] ************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Create /usr/share/ansible directory] *************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install collections from Ansible galaxy] *********************************",
                                "changed: [testbed-manager] => (item=ansible.netcommon)",
                                "changed: [testbed-manager] => (item=ansible.posix)",
                                "changed: [testbed-manager] => (item=community.docker>=3.10.2)",
                                "",
                                "TASK [Install local collections] ***********************************************",
                                "changed: [testbed-manager] => (item=ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=ansible-collection-services)",
                                "",
                                "PLAY [Create operator user] ****************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "changed: [testbed-manager] => (item=adm)",
                                "changed: [testbed-manager] => (item=sudo)",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LANG=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8)",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Recursively change ownership of /opt/venv] *******************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000014-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:46:18.319304Z",
                            "start": "2026-03-26T01:40:03.063270Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000014",
                        "name": "Run manager part 0"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "It is now already possible to log in to the manager with 'make login'."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:46:18.373416Z",
                            "start": "2026-03-26T01:46:18.331744Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000015",
                        "name": "Point out that the log in on the manager is now possible"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:46:18.416063Z",
                            "start": "2026-03-26T01:46:18.380570Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000016",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  -e manager_version=9.5.0 \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                            "delta": "0:02:38.999822",
                            "end": "2026-03-26 01:48:57.739141",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  -e manager_version=9.5.0 \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000017-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:46:18.739319",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Run manager part 1] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nincluded: /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nskipping: [testbed-manager]\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Update package cache] *************************\nchanged: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Copy SSH public key] *****************************************************\nchanged: [testbed-manager]\n\nTASK [Copy SSH private key] ****************************************************\nchanged: [testbed-manager]\n\nTASK [Create configuration directory] ******************************************\nchanged: [testbed-manager]\n\nTASK [Copy testbed repo] *******************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager]\n\nTASK [Install python requirements in venv] *************************************\nok: [testbed-manager] => (item=Jinja2)\nok: [testbed-manager] => (item=PyYAML)\nok: [testbed-manager] => (item=packaging)\nchanged: [testbed-manager] => (item=python-gilt==1.2.3)\nok: [testbed-manager] => (item=requests>=2.32.2)\nok: [testbed-manager] => (item=docker>=7.1.0)\n\nTASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************\nchanged: [testbed-manager]\n\nTASK [Copy testbed custom CA certificate on CentOS] ****************************\nskipping: [testbed-manager]\n\nTASK [Run update-ca-certificates on Debian/Ubuntu] *****************************\nchanged: [testbed-manager]\n\nTASK [Run update-ca-trust on RedHat] *******************************************\nskipping: [testbed-manager]\n\nTASK [Run manager part 2] ******************************************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nok: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Run manager part 1] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "included: /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Copy SSH public key] *****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy SSH private key] ****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create configuration directory] ******************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed repo] *******************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install python requirements in venv] *************************************",
                                "ok: [testbed-manager] => (item=Jinja2)",
                                "ok: [testbed-manager] => (item=PyYAML)",
                                "ok: [testbed-manager] => (item=packaging)",
                                "changed: [testbed-manager] => (item=python-gilt==1.2.3)",
                                "ok: [testbed-manager] => (item=requests>=2.32.2)",
                                "ok: [testbed-manager] => (item=docker>=7.1.0)",
                                "",
                                "TASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed custom CA certificate on CentOS] ****************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-certificates on Debian/Ubuntu] *****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-trust on RedHat] *******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run manager part 2] ******************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000017-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:48:58.069901Z",
                            "start": "2026-03-26T01:46:18.449122Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000017",
                        "name": "Run manager part 1 + 2"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "sudo",
                                "shutdown",
                                "-r",
                                "now"
                            ],
                            "delta": "0:00:00.978791",
                            "end": "2026-03-26 01:48:59.380492",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 sudo shutdown -r now",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000018-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:48:58.401701",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-000000000018-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:48:59.629062Z",
                            "start": "2026-03-26T01:48:58.117065Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000018",
                        "name": "Reboot manager"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 16,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.54",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:49:16.056597Z",
                            "start": "2026-03-26T01:48:59.644674Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000019",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-26 01:49:16.087803",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-26 01:50:16.090365",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T01:50:16.115180Z",
                            "start": "2026-03-26T01:49:16.064059Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001a",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "/opt/configuration/scripts/deploy-manager.sh"
                            ],
                            "delta": "0:25:05.319279",
                            "end": "2026-03-26 02:15:22.182910",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 /opt/configuration/scripts/deploy-manager.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 01:50:16.863631",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY MANAGER\n\n+ set -e\n+ echo\n+ echo '# DEPLOY MANAGER'\n+ echo\n+ cat /opt/manager-vars.sh\nexport NUMBER_OF_NODES=6\n\nexport CEPH_VERSION=reef\nexport CONFIGURATION_VERSION=main\nexport MANAGER_VERSION=9.5.0\nexport OPENSTACK_VERSION=2024.2\n\nexport ARA=false\nexport DEPLOY_MODE=manager\nexport TEMPEST=false\nexport IS_ZUUL=true\n\nexport MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\nexport EXTERNAL_API=false\n\nexport IMAGE_USER=ubuntu\nexport IMAGE_NODE_USER=ubuntu\n\nexport CEPH_STACK=ceph-ansible\n\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver\n+ docker version\nClient: Docker Engine - Community\n Version:           27.5.1\n API version:       1.47\n Go version:        go1.22.11\n Git commit:        9f9e405\n Built:             Wed Jan 22 13:41:48 2025\n OS/Arch:           linux/amd64\n Context:           default\n\nServer: Docker Engine - Community\n Engine:\n  Version:          27.5.1\n  API version:      1.47 (minimum version 1.24)\n  Go version:       go1.22.11\n  Git commit:       4c9b3b0\n  Built:            Wed Jan 22 13:41:48 2025\n  OS/Arch:          linux/amd64\n  Experimental:     false\n containerd:\n  Version:          v2.2.2\n  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9\n runc:\n  Version:          1.3.4\n  GitCommit:        v1.3.4-0-gd6d73eb8\n docker-init:\n  Version:          0.19.0\n  GitCommit:        de40ad0\n+ sh -c /opt/configuration/scripts/deploy/000-manager.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-manager-version.sh 9.5.0\n+ set -e\n+ VERSION=9.5.0\n+ sed -i 's/manager_version: .*/manager_version: 9.5.0/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]\n+ sed -i /ceph_version:/d /opt/configuration/environments/manager/configuration.yml\n+ sed -i /openstack_version:/d /opt/configuration/environments/manager/configuration.yml\n+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh\n/opt/configuration ~\n+ set -e\n+ pushd /opt/configuration\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging\nRequirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)\nRequirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)\nRequirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)\nRequirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)\nRequirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)\nRequirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)\nRequirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)\nRequirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)\nRequirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)\nRequirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)\nRequirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)\nRequirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)\nRequirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)\nRequirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)\n++ which gilt\n+ GILT=/opt/venv/bin/gilt\n+ /opt/venv/bin/gilt overlay\nosism.cfg-generics:\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/\n  - running `rm render-images.py` in /opt/configuration/environments/manager/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/\n  - running `rm set-versions.py` in /opt/configuration/environments/\n~\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ popd\n+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]\n+ [[ ceph-ansible == \\r\\o\\o\\k ]]\n++ semver 9.5.0 7.0.0\n+ [[ 1 -ge 0 ]]\n+ echo 'enable_osism_kubernetes: true'\n++ semver 9.5.0 10.0.0-0\n+ [[ -1 -ge 0 ]]\n++ semver 2024.2 2025.1\n+ [[ -1 -ge 0 ]]\n+ /opt/configuration/scripts/enable-resource-nodes.sh\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml\n\nPLAY [Copy custom facts] *******************************************************\n\nTASK [Create custom facts directory] *******************************************\nok: [testbed-manager]\n\nTASK [Copy fact files] *********************************************************\nchanged: [testbed-manager]\n\nPLAY [Before the deployment of the manager] ************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get /opt/manager-vars.sh] ************************************************\nok: [testbed-manager]\n\nTASK [Add ara_server_mariadb_volume_type parameter] ****************************\nchanged: [testbed-manager]\n\nTASK [Add netbox_enable parameter] *********************************************\nskipping: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Check if /etc/OTC_region exist] ******************************************\nok: [testbed-manager]\n\nTASK [Add nova_compute_virt_type parameter] ************************************\nskipping: [testbed-manager]\n\nPLAY [Apply role traefik] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply traefik role] ******************************************************\nincluded: osism.services.traefik for testbed-manager\n\nTASK [osism.services.traefik : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager\n\nTASK [osism.services.traefik : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/traefik)\nchanged: [testbed-manager] => (item=/opt/traefik/certificates)\nchanged: [testbed-manager] => (item=/opt/traefik/configuration)\n\nTASK [osism.services.traefik : Copy configuration files] ***********************\nchanged: [testbed-manager] => (item=traefik.yml)\nchanged: [testbed-manager] => (item=traefik.env)\nchanged: [testbed-manager] => (item=certificates.yml)\n\nTASK [osism.services.traefik : Copy certificate cert files] ********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy certificate key files] *********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy dynamic configuration] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.traefik : Remove dynamic configuration] *******************\nok: [testbed-manager]\n\nTASK [osism.services.traefik : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager\n\nTASK [osism.services.traefik : Create traefik external network] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Manage traefik service] *************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.traefik : Restart traefik service] *************\nskipping: [testbed-manager]\n\nPLAY [Deploy manager service] **************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply manager role] ******************************************************\nincluded: osism.services.manager for testbed-manager\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/ansible)\nchanged: [testbed-manager] => (item=/opt/archive)\nchanged: [testbed-manager] => (item=/opt/manager/configuration)\nchanged: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nchanged: [testbed-manager] => (item=/opt/manager/secrets)\nchanged: [testbed-manager] => (item=/opt/ansible/secrets)\nchanged: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nchanged: [testbed-manager] => (item=ara)\nchanged: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nchanged: [testbed-manager] => (item=conductor)\nchanged: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nchanged: [testbed-manager] => (item=osism)\nchanged: [testbed-manager] => (item=osism-update-docker)\nchanged: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nchanged: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nok: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nskipping: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:0.20251130.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:0.20251130.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:0.20251130.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:0.20251130.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:0.20251130.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:0.20251130.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:0.20251130.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:0.20251130.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:0.20251130.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:0.20251130.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   \n\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                         COMMAND                  SERVICE                CREATED         STATUS                        PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:0.20251130.0           \"/entrypoint.sh osis\u2026\"   ceph-ansible           2 minutes ago   Up About a minute (healthy)   \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:0.20251130.0          \"/entrypoint.sh osis\u2026\"   kolla-ansible          2 minutes ago   Up About a minute (healthy)   \nmanager-api-1                    registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago   Up 2 minutes (healthy)        192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                    \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago   Up 2 minutes (healthy)        8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago   Up 2 minutes (healthy)        \nmanager-flower-1                 registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago   Up 2 minutes (healthy)        \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:0.20251130.0   \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   2 minutes ago   Up About a minute (healthy)   \nmanager-listener-1               registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago   Up 2 minutes (healthy)        \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4          \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago   Up 2 minutes (healthy)        3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago   Up 2 minutes (healthy)        \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine      \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago   Up 2 minutes (healthy)        6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:0.20251130.0          \"/entrypoint.sh osis\u2026\"   osism-ansible          2 minutes ago   Up About a minute (healthy)   \nosism-frontend                   registry.osism.tech/osism/osism-frontend:0.20251130.1         \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago   Up 2 minutes                  192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:0.20251130.0       \"/entrypoint.sh osis\u2026\"   osism-kubernetes       2 minutes ago   Up About a minute (healthy)   \nosismclient                      registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago   Up 2 minutes (healthy)        \n++ semver 9.5.0 7.0.0\n+ [[ 1 -ge 0 ]]\n+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg\n+ osism apply resolvconf -l testbed-manager\n\u001b[32m2026-03-26 01:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b338eb90-9c5b-4feb-973b-e917f7311fb8 (resolvconf) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b338eb90-9c5b-4feb-973b-e917f7311fb8 (resolvconf) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role resolvconf] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nThursday 26 March 2026  01:55:17 +0000 (0:00:00.157)       0:00:00.157 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nThursday 26 March 2026  01:55:21 +0000 (0:00:03.974)       0:00:04.131 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nThursday 26 March 2026  01:55:21 +0000 (0:00:00.071)       0:00:04.203 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nThursday 26 March 2026  01:55:21 +0000 (0:00:00.091)       0:00:04.294 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nThursday 26 March 2026  01:55:21 +0000 (0:00:00.081)       0:00:04.376 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nThursday 26 March 2026  01:55:22 +0000 (0:00:01.247)       0:00:05.623 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nThursday 26 March 2026  01:55:22 +0000 (0:00:00.066)       0:00:05.690 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nThursday 26 March 2026  01:55:23 +0000 (0:00:00.528)       0:00:06.218 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nThursday 26 March 2026  01:55:23 +0000 (0:00:00.091)       0:00:06.310 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nThursday 26 March 2026  01:55:23 +0000 (0:00:00.562)       0:00:06.872 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nThursday 26 March 2026  01:55:25 +0000 (0:00:01.131)       0:00:08.004 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nThursday 26 March 2026  01:55:26 +0000 (0:00:01.017)       0:00:09.021 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nThursday 26 March 2026  01:55:26 +0000 (0:00:00.084)       0:00:09.106 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  01:55:27 +0000 (0:00:01.180)       0:00:10.286 ******** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.97s\nosism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.25s\nosism.commons.resolvconf : Restart systemd-resolved service ------------- 1.18s\nosism.commons.resolvconf : Copy configuration files --------------------- 1.13s\nosism.commons.resolvconf : Start/enable systemd-resolved service -------- 1.02s\nosism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.56s\nosism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.53s\nosism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.09s\nosism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s\nosism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s\nosism.commons.resolvconf : Include distribution specific installation tasks --- 0.08s\nosism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s\nosism.commons.resolvconf : Install package systemd-resolved ------------- 0.07s\n+ osism apply sshconfig\n\u001b[32m2026-03-26 01:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c0f329a2-f521-4b7e-839d-4cec16d416a2 (sshconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c0f329a2-f521-4b7e-839d-4cec16d416a2 (sshconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role sshconfig] ****************************************************\n\nTASK [osism.commons.sshconfig : Get home directory of operator user] ***********\nThursday 26 March 2026  01:55:44 +0000 (0:00:00.169)       0:00:00.169 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************\nThursday 26 March 2026  01:55:44 +0000 (0:00:00.551)       0:00:00.721 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure config for each host exist] *************\nThursday 26 March 2026  01:55:45 +0000 (0:00:00.563)       0:00:01.284 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\nTASK [osism.commons.sshconfig : Add extra config] ******************************\nThursday 26 March 2026  01:55:51 +0000 (0:00:06.025)       0:00:07.310 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Assemble ssh config] ***************************\nThursday 26 March 2026  01:55:51 +0000 (0:00:00.088)       0:00:07.398 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  01:55:52 +0000 (0:00:00.631)       0:00:08.029 ******** \n=============================================================================== \nosism.commons.sshconfig : Ensure config for each host exist ------------- 6.03s\nosism.commons.sshconfig : Assemble ssh config --------------------------- 0.63s\nosism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.56s\nosism.commons.sshconfig : Get home directory of operator user ----------- 0.55s\nosism.commons.sshconfig : Add extra config ------------------------------ 0.09s\n+ osism apply known-hosts\n\u001b[32m2026-03-26 01:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1fa85f69-74d9-41fc-88eb-077113f41686 (known-hosts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1fa85f69-74d9-41fc-88eb-077113f41686 (known-hosts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role known_hosts] **************************************************\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***\nThursday 26 March 2026  01:56:09 +0000 (0:00:00.179)       0:00:00.179 ******** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***\nThursday 26 March 2026  01:56:15 +0000 (0:00:06.471)       0:00:06.651 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:15 +0000 (0:00:00.175)       0:00:06.827 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJ02fclj3yBlaRFCMRlBqzxa6JpoNn0+0cesF8DzED/Y)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCjl441PvBESBlyBRx1uu1H4fE7tRu8wdi33Q8JFqeLTaQZjzDEdIz6uCS4P7vly4E/qLLf8bB/8UcNRyQNcICQGP4n1FNR/EKvBi3/Gyc8AFEHmafJBMQ+fPJKZT16DdkyJaug5SebtzLWoBu55VpcJBQEuazelbRRNG8Xz6iBrkfu/Kec/g0Rno6KOcdUjXqp9eXijo2MLaMUS8xKb9xvjzA8lG94Nfv5kLJxesAhlOo3+QdD7zaeMZ7NUtCfAV0/o8Ruvk/OCfm+9xqBh34HvSOpZATYXPchqL5K63v6jn0PFjt3YFZ0gIUmA1WSFQpb6laHtVAv7J1OW2xdYo6gotU1Cr37/4ig+HHn9P+oQgsmR48E/LQGDOE8e2lb9vbBc00XUT6BPP5pLf9JyRJiE8oeTuYbPu7EAcxRe0Ub7IiJPGMHrzRYtxLMNbIcIC3Albt9+yYod/l1CB/xtv2cBu/ChuADP3Zn4908Xfv7x+enwcnvCi1DiX8TxkEolms=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBP5uNsSXygIZlJfKHbiEnkj18urKp5QMG+27OXFSyVeZCLFzqulV6Rzhv/iPWmiLWxR1gs0vX6AnWLfHyOAmsI4=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:17 +0000 (0:00:01.243)       0:00:08.071 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC7MP4ethAPLioYMXHsztWTZ5ixht+3Sq7KMlb6HBjt4WWGu5avxBzCqnNkYFhF7TM+NZp4FPbph0s1vt1MbCjwAX101M9p/xXyIg6HqiULDQGPnSM9m+LRiEsrASOQGF6aFihMRdcJRhnREJZq9qGdmItuWj9KhGkPm8uRtSFbq+buvlAumNVyRGX/2g9hRT6OPPaBs/0vnzqWm0PY24uelp14behZeTqbe67Xef6Ejb2CPSpASVYAK7cBu4QLBtZt+ComHffq2pp5kC/BZzYHIkNSsmHez8A+uBRCSgkLTqIoCB1KlEDd4YUFkUqRwLSO3P+YeXwrdsuLV64aUMYp+mx/hn1UhZf52iPN5/4dhu/u1pNdyfBgkAI8YresJOdSH6Ztn91oOS0NfNvirYGQGgZQ8F6GAHw/Jqj4Tsu0/0pBD6p34E7ve6hpBikPhSLPtsJmfT5TuU1FdhBMtq6IGmYZfCa24W3TuSRsZnpNeoagjyEJOIujgECw1Gl5Vfc=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEJBAfKNBqc6hZaPCToSyiH071v/szTHinZTYmvWtQiYY2BD9OmGZbDU8lB1KJXImA4sH06NV8BnsqYFdWToJUQ=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHe8Qmub5XmBttmFEUBkU1KnPXMx15s5oXemfYKWwqe/)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:18 +0000 (0:00:01.203)       0:00:09.274 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+4xLf8E9LJe8aia/EpM0v0rX+8sflF8vLZvkHgEXWEhHuSqtxzIL8ACbdmnHD33O2W0UBsUAyKb19kCZTvLXTqmZ2+J24oE3ZFXc8YL5hTd5Yl4TRhBUjg9m31EizACs/0jAMW2ZKSmiAC0t9dhTj2ZMMJ9YMi/kGEhZ1c3/HNB0H3V8krwQERPfDvu8cEHaZVKLA2E86hUr7O4LpXVaMjYarBwX2WTY5YK2l3fkb7rJAigD7P858LeetHWGn6w4gw6illPyvLV50RMhF4BD/KdhvdPV2Zilyh/H23zhuvOy1he82obLlwFNK7mibsaxgEJrsx0VpcFsA6bygql94UvQR3dKTb4Rc34++oxIQgdFsdrIKHadlhGcaEekjU6ZI0BKd0qE511yymd1N3ooYoVHbf8mhhr1oKWq/os46wBLhEyuyNYD3xS2L/YX2jFpKgiU4PIYyJZy+WDIkiuZNjucyKTdIci0+wIJYghYdt9ZWtDIyBEJn6QjDxRRYWaE=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLOteKbqWmsecSgYs+xqbb6nNATEUQK4q13kbcpg1b1i+s0qETmOp+d8ADubOahpy9BafeOlLLvip9rzRVjYfSs=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICb08xQfqtFt6WLy7uNMC8Hgb7gIH7sGbdc6Gp7j5cy2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:19 +0000 (0:00:01.208)       0:00:10.482 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDM4uc4YzFGwjqMlk336TEFW9Sr5NydBFeDixkdguq0xoKZ7Av9HzodJRVzGN627myZhV4E815uWI3x9reyYZn0RcrcDzleefzPfPoWxct0zKx0MfT2zwMJ22JNwOTn6REmWHbevXA5ITeutZAPxJdPrwoS8s1a1Buj3Ky0tfIaCqCRAkIV5P1pJRFSlt7gZL0ImYnlZoIDfKZ/cBT9BzsRT8CWHHY37KfeVwJm37huH9rMBh/eHolG5yLNXGvKrTq0QKC1mMY2qK0APt4GmFl8KPTnNLajycV5GheWblGibpQGFM93+2Z/WqnRJf82EIxu8yCqPo3/1FhS5lFWtQRK9+I/5G7KpU3S/WCcAmUn+w2yg1uU3CAEV+RycYxvp2lqJTbwlZx/pTBotJ9R7cUBQ5aW0Nsvp8aytFY0HRc4Xa8z0Pg6UlJTSdG0IENfARZSKYkJ1g+2JRpe/lOoXOqpWEA4KW+XdPodGxf0D2YqCd/rGxlBauwYYWbZ3Imo7ic=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAubqEkG0vhsyENkRhNUdy8+oCOQJ2VjqsD1HLNi3GC9S06/AsvC5szSrOM8xPq6YUCb0ZAm8wLM6tpF+FpQCkI=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIILAw9GUy2vpfwElSJ1fGmtyFuqmYW2r9XbGM4GRo7eY)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:20 +0000 (0:00:01.216)       0:00:11.699 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ5PAiQENk+ItZkO7lCzckgIds0Q2WgCbLVLzPfl9J/MU02VMGGoMDDgddubcewkixXkbNvIDK1stUBpmzi0L0j6dHcUMdFvRgqm8GCoMsXNrfHW/ct+73eNQxULwT9bQllXOZr2nB7MltedCplvwjiFJBdmFrGAIp+GrXL0PMApIE/dd1f1qPoBloRBPpdJX1CjF1OGaQsZ4qpz7WpJrlw7thtAHdcylkEbcstauBnNgTubYjVaIo2uIJXiCWNIL0fPhNIHAMKddURALWDzhs+Xh/MW65iFiI8YfPERzwYogm9ZGXr3rX08YWezcrnK5zImuNAFUMNYrmHbLAjzsIRCewlKYooBB2dtd3+XDSqNrvX7+tVW6K8NlU917X/WU52YVtcXFisGbVrw/elyi1iLd7EwmJ9QCUMobcLT0ImLsPJAR5WpjsxApgqPjXLiRWoBFIT4ZJh3VGUe5BPDf8dT/Ij1IENkwR4wNsSm+8yKbokryeCpNJ/83tRYElNd8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKbUTEgC6bw2kPbQ69BQKKbddt6FzzQXjl7gPSrh85KeKra/4AOp2yhnEkwPsOgQiMMH5NRrWQXsymB0y+9b1Q8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGuDGp4QP0xBuwhTX5o6OTk2Lr58UqgVCuxQtBHOizHi)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:21 +0000 (0:00:01.175)       0:00:12.875 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNUp+hESk7uzg35hbkvvf+NUU3AkgbCjVB2Zkr8YJaR+BFh9GPPoZZsfxubkFFdXWW3nEN+8mFY1PIXNSyE0jYU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCtvM026ThskiZdDILB9UjPU47PsDHtiHncBI7eSm7qoZ/vFmGYxPZa4NYpRQd13zEH+bhVZEQoiEJjSXZTdefjqImCiIVEtQYLak7OohLyh+btJqEp4MZX1p+Ef87NJt6iZXTmRO9hJa7iuKN5UU6QTJuvgvy/KEw7KR/yqIwc90ds6MtpKtAKEnqodONOccyVKbdgfAngkmZ2uye4IM8jrwODbjTbvmztDxJ3Sp24S7z173HSnLssG6zn5hxZxlMJQPsshb1LZB7JEF6lISzJIVNgyjszZc+PJPIXOQ0eOfc5gVES9pBSljtremJUbyNhJTxEYrER41dO5QupDhOo2kWeNPdHCnP2TzghsdM8MYFFRQrxKuPY9/o9ba8oW0xMHcVSQiUXpxZ0envlNeCzmYATs6z4y/XPHjA1K+boWRjzlBpTOsPyxFYR8NWg16MV5V4YwxIGBkMwr86umV4mJ+4qP03CktNELRx8hMDTGQ1vt5H3sjQnWKAb5bwbJnk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC/uIeNDlNnWqqbf4nQ1iFXm8xqmTno9JV+7U0mLmoBx)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:22 +0000 (0:00:01.136)       0:00:14.011 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCgKIZ+Izl5D9pHv6Kxhr5AyZEEKn9/BK9MNhQb/QhaO78ATvWpiJlyn30Si4IfRm0v7dCgB0sW8NB5CvAGq/ZGVQE3hr4tNBZrraVdJW4ItePEAsUdTBKwaGp9SU7bwlkZDWNr4EKmSRIefON4mC+Ybn4LJccP5VfAIqMv7w/fcREx9xmgnP7OprQLP3T35T1ObQBXrDu4nOSoxOUa3eeTBzb0M2JXK31R8MAkfBpl3xt/3e/97jB5oOcK9t9kbC3QwZmIYAcx+XCSmxyQ7kcR5PTtZoBVwzPXqPMU0tCc0UXDEJRnsTrXIJ0wArclmDuuIyLundoxRn57nEmHByjdHzJ5u/RqZYNJHs0LWyPDC7FOQ2ek+xnsnzjizES2z96sVFk385BUNTp+/Iv74/8jOP7Glx2+Q3TQc5BCM45SISAWv/C65y24IJLwG5hfzk0+g1I/moUPPDThzsC9weUzFs4C0dDIYlckw90A0PCUIFSGzaLQtti2t3GTfb6oISU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK2OXpbGe9PH2FJGGvIa5N+853ic5OP4B+gsXUDRcJHyzkcI+mMK2k9rDDpbgobH0QeLp+WUIjUIHSmvkdzCtC4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIK2b9SZZ/SJ6wJY/vZIQL6KvWIfJRQxK2nxhYfFkytaC)\u001b[0m\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***\nThursday 26 March 2026  01:56:24 +0000 (0:00:01.211)       0:00:15.223 ******** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***\nThursday 26 March 2026  01:56:29 +0000 (0:00:05.758)       0:00:20.981 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:30 +0000 (0:00:00.184)       0:00:21.166 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCjl441PvBESBlyBRx1uu1H4fE7tRu8wdi33Q8JFqeLTaQZjzDEdIz6uCS4P7vly4E/qLLf8bB/8UcNRyQNcICQGP4n1FNR/EKvBi3/Gyc8AFEHmafJBMQ+fPJKZT16DdkyJaug5SebtzLWoBu55VpcJBQEuazelbRRNG8Xz6iBrkfu/Kec/g0Rno6KOcdUjXqp9eXijo2MLaMUS8xKb9xvjzA8lG94Nfv5kLJxesAhlOo3+QdD7zaeMZ7NUtCfAV0/o8Ruvk/OCfm+9xqBh34HvSOpZATYXPchqL5K63v6jn0PFjt3YFZ0gIUmA1WSFQpb6laHtVAv7J1OW2xdYo6gotU1Cr37/4ig+HHn9P+oQgsmR48E/LQGDOE8e2lb9vbBc00XUT6BPP5pLf9JyRJiE8oeTuYbPu7EAcxRe0Ub7IiJPGMHrzRYtxLMNbIcIC3Albt9+yYod/l1CB/xtv2cBu/ChuADP3Zn4908Xfv7x+enwcnvCi1DiX8TxkEolms=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBP5uNsSXygIZlJfKHbiEnkj18urKp5QMG+27OXFSyVeZCLFzqulV6Rzhv/iPWmiLWxR1gs0vX6AnWLfHyOAmsI4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJ02fclj3yBlaRFCMRlBqzxa6JpoNn0+0cesF8DzED/Y)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:31 +0000 (0:00:01.113)       0:00:22.279 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC7MP4ethAPLioYMXHsztWTZ5ixht+3Sq7KMlb6HBjt4WWGu5avxBzCqnNkYFhF7TM+NZp4FPbph0s1vt1MbCjwAX101M9p/xXyIg6HqiULDQGPnSM9m+LRiEsrASOQGF6aFihMRdcJRhnREJZq9qGdmItuWj9KhGkPm8uRtSFbq+buvlAumNVyRGX/2g9hRT6OPPaBs/0vnzqWm0PY24uelp14behZeTqbe67Xef6Ejb2CPSpASVYAK7cBu4QLBtZt+ComHffq2pp5kC/BZzYHIkNSsmHez8A+uBRCSgkLTqIoCB1KlEDd4YUFkUqRwLSO3P+YeXwrdsuLV64aUMYp+mx/hn1UhZf52iPN5/4dhu/u1pNdyfBgkAI8YresJOdSH6Ztn91oOS0NfNvirYGQGgZQ8F6GAHw/Jqj4Tsu0/0pBD6p34E7ve6hpBikPhSLPtsJmfT5TuU1FdhBMtq6IGmYZfCa24W3TuSRsZnpNeoagjyEJOIujgECw1Gl5Vfc=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEJBAfKNBqc6hZaPCToSyiH071v/szTHinZTYmvWtQiYY2BD9OmGZbDU8lB1KJXImA4sH06NV8BnsqYFdWToJUQ=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHe8Qmub5XmBttmFEUBkU1KnPXMx15s5oXemfYKWwqe/)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:32 +0000 (0:00:01.189)       0:00:23.469 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+4xLf8E9LJe8aia/EpM0v0rX+8sflF8vLZvkHgEXWEhHuSqtxzIL8ACbdmnHD33O2W0UBsUAyKb19kCZTvLXTqmZ2+J24oE3ZFXc8YL5hTd5Yl4TRhBUjg9m31EizACs/0jAMW2ZKSmiAC0t9dhTj2ZMMJ9YMi/kGEhZ1c3/HNB0H3V8krwQERPfDvu8cEHaZVKLA2E86hUr7O4LpXVaMjYarBwX2WTY5YK2l3fkb7rJAigD7P858LeetHWGn6w4gw6illPyvLV50RMhF4BD/KdhvdPV2Zilyh/H23zhuvOy1he82obLlwFNK7mibsaxgEJrsx0VpcFsA6bygql94UvQR3dKTb4Rc34++oxIQgdFsdrIKHadlhGcaEekjU6ZI0BKd0qE511yymd1N3ooYoVHbf8mhhr1oKWq/os46wBLhEyuyNYD3xS2L/YX2jFpKgiU4PIYyJZy+WDIkiuZNjucyKTdIci0+wIJYghYdt9ZWtDIyBEJn6QjDxRRYWaE=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLOteKbqWmsecSgYs+xqbb6nNATEUQK4q13kbcpg1b1i+s0qETmOp+d8ADubOahpy9BafeOlLLvip9rzRVjYfSs=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICb08xQfqtFt6WLy7uNMC8Hgb7gIH7sGbdc6Gp7j5cy2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:33 +0000 (0:00:01.183)       0:00:24.653 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAubqEkG0vhsyENkRhNUdy8+oCOQJ2VjqsD1HLNi3GC9S06/AsvC5szSrOM8xPq6YUCb0ZAm8wLM6tpF+FpQCkI=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIILAw9GUy2vpfwElSJ1fGmtyFuqmYW2r9XbGM4GRo7eY)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDM4uc4YzFGwjqMlk336TEFW9Sr5NydBFeDixkdguq0xoKZ7Av9HzodJRVzGN627myZhV4E815uWI3x9reyYZn0RcrcDzleefzPfPoWxct0zKx0MfT2zwMJ22JNwOTn6REmWHbevXA5ITeutZAPxJdPrwoS8s1a1Buj3Ky0tfIaCqCRAkIV5P1pJRFSlt7gZL0ImYnlZoIDfKZ/cBT9BzsRT8CWHHY37KfeVwJm37huH9rMBh/eHolG5yLNXGvKrTq0QKC1mMY2qK0APt4GmFl8KPTnNLajycV5GheWblGibpQGFM93+2Z/WqnRJf82EIxu8yCqPo3/1FhS5lFWtQRK9+I/5G7KpU3S/WCcAmUn+w2yg1uU3CAEV+RycYxvp2lqJTbwlZx/pTBotJ9R7cUBQ5aW0Nsvp8aytFY0HRc4Xa8z0Pg6UlJTSdG0IENfARZSKYkJ1g+2JRpe/lOoXOqpWEA4KW+XdPodGxf0D2YqCd/rGxlBauwYYWbZ3Imo7ic=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:34 +0000 (0:00:01.177)       0:00:25.830 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKbUTEgC6bw2kPbQ69BQKKbddt6FzzQXjl7gPSrh85KeKra/4AOp2yhnEkwPsOgQiMMH5NRrWQXsymB0y+9b1Q8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGuDGp4QP0xBuwhTX5o6OTk2Lr58UqgVCuxQtBHOizHi)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ5PAiQENk+ItZkO7lCzckgIds0Q2WgCbLVLzPfl9J/MU02VMGGoMDDgddubcewkixXkbNvIDK1stUBpmzi0L0j6dHcUMdFvRgqm8GCoMsXNrfHW/ct+73eNQxULwT9bQllXOZr2nB7MltedCplvwjiFJBdmFrGAIp+GrXL0PMApIE/dd1f1qPoBloRBPpdJX1CjF1OGaQsZ4qpz7WpJrlw7thtAHdcylkEbcstauBnNgTubYjVaIo2uIJXiCWNIL0fPhNIHAMKddURALWDzhs+Xh/MW65iFiI8YfPERzwYogm9ZGXr3rX08YWezcrnK5zImuNAFUMNYrmHbLAjzsIRCewlKYooBB2dtd3+XDSqNrvX7+tVW6K8NlU917X/WU52YVtcXFisGbVrw/elyi1iLd7EwmJ9QCUMobcLT0ImLsPJAR5WpjsxApgqPjXLiRWoBFIT4ZJh3VGUe5BPDf8dT/Ij1IENkwR4wNsSm+8yKbokryeCpNJ/83tRYElNd8=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:35 +0000 (0:00:01.199)       0:00:27.030 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCtvM026ThskiZdDILB9UjPU47PsDHtiHncBI7eSm7qoZ/vFmGYxPZa4NYpRQd13zEH+bhVZEQoiEJjSXZTdefjqImCiIVEtQYLak7OohLyh+btJqEp4MZX1p+Ef87NJt6iZXTmRO9hJa7iuKN5UU6QTJuvgvy/KEw7KR/yqIwc90ds6MtpKtAKEnqodONOccyVKbdgfAngkmZ2uye4IM8jrwODbjTbvmztDxJ3Sp24S7z173HSnLssG6zn5hxZxlMJQPsshb1LZB7JEF6lISzJIVNgyjszZc+PJPIXOQ0eOfc5gVES9pBSljtremJUbyNhJTxEYrER41dO5QupDhOo2kWeNPdHCnP2TzghsdM8MYFFRQrxKuPY9/o9ba8oW0xMHcVSQiUXpxZ0envlNeCzmYATs6z4y/XPHjA1K+boWRjzlBpTOsPyxFYR8NWg16MV5V4YwxIGBkMwr86umV4mJ+4qP03CktNELRx8hMDTGQ1vt5H3sjQnWKAb5bwbJnk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNUp+hESk7uzg35hbkvvf+NUU3AkgbCjVB2Zkr8YJaR+BFh9GPPoZZsfxubkFFdXWW3nEN+8mFY1PIXNSyE0jYU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC/uIeNDlNnWqqbf4nQ1iFXm8xqmTno9JV+7U0mLmoBx)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nThursday 26 March 2026  01:56:37 +0000 (0:00:01.197)       0:00:28.228 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCgKIZ+Izl5D9pHv6Kxhr5AyZEEKn9/BK9MNhQb/QhaO78ATvWpiJlyn30Si4IfRm0v7dCgB0sW8NB5CvAGq/ZGVQE3hr4tNBZrraVdJW4ItePEAsUdTBKwaGp9SU7bwlkZDWNr4EKmSRIefON4mC+Ybn4LJccP5VfAIqMv7w/fcREx9xmgnP7OprQLP3T35T1ObQBXrDu4nOSoxOUa3eeTBzb0M2JXK31R8MAkfBpl3xt/3e/97jB5oOcK9t9kbC3QwZmIYAcx+XCSmxyQ7kcR5PTtZoBVwzPXqPMU0tCc0UXDEJRnsTrXIJ0wArclmDuuIyLundoxRn57nEmHByjdHzJ5u/RqZYNJHs0LWyPDC7FOQ2ek+xnsnzjizES2z96sVFk385BUNTp+/Iv74/8jOP7Glx2+Q3TQc5BCM45SISAWv/C65y24IJLwG5hfzk0+g1I/moUPPDThzsC9weUzFs4C0dDIYlckw90A0PCUIFSGzaLQtti2t3GTfb6oISU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK2OXpbGe9PH2FJGGvIa5N+853ic5OP4B+gsXUDRcJHyzkcI+mMK2k9rDDpbgobH0QeLp+WUIjUIHSmvkdzCtC4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIK2b9SZZ/SJ6wJY/vZIQL6KvWIfJRQxK2nxhYfFkytaC)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write static known_hosts entries] ************\nThursday 26 March 2026  01:56:38 +0000 (0:00:01.261)       0:00:29.489 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Write extra known_hosts entries] *************\nThursday 26 March 2026  01:56:38 +0000 (0:00:00.166)       0:00:29.656 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Delete known_hosts entries] ******************\nThursday 26 March 2026  01:56:38 +0000 (0:00:00.071)       0:00:29.728 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Set file permissions] ************************\nThursday 26 March 2026  01:56:38 +0000 (0:00:00.066)       0:00:29.794 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  01:56:39 +0000 (0:00:00.807)       0:00:30.602 ******** \n=============================================================================== \nosism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 6.47s\nosism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.76s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.26s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.24s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.22s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.21s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.21s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.19s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.14s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.11s\nosism.commons.known_hosts : Set file permissions ------------------------ 0.81s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.18s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.18s\nosism.commons.known_hosts : Write static known_hosts entries ------------ 0.17s\n+ osism apply squid\n\u001b[32m2026-03-26 01:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 22181c16-c6bd-4755-9a52-705156a3718e (squid) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 22181c16-c6bd-4755-9a52-705156a3718e (squid) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role squid] ********************************************************\n\nTASK [osism.services.squid : Include install tasks] ****************************\nThursday 26 March 2026  01:56:56 +0000 (0:00:00.184)       0:00:00.184 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.squid : Install required packages] ************************\nThursday 26 March 2026  01:56:56 +0000 (0:00:00.108)       0:00:00.293 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Create required directories] **********************\nThursday 26 March 2026  01:56:58 +0000 (0:00:01.915)       0:00:02.208 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m\n\nTASK [osism.services.squid : Copy squid configuration files] *******************\nThursday 26 March 2026  01:57:00 +0000 (0:00:01.319)       0:00:03.528 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m\n\nTASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***\nThursday 26 March 2026  01:57:01 +0000 (0:00:01.148)       0:00:04.676 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Copy docker-compose.yml file] *********************\nThursday 26 March 2026  01:57:01 +0000 (0:00:00.386)       0:00:05.062 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Manage squid service] *****************************\nThursday 26 March 2026  01:57:02 +0000 (0:00:00.961)       0:00:06.023 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Restart squid service] *****************\nThursday 26 March 2026  01:57:39 +0000 (0:00:36.629)       0:00:42.653 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******\nThursday 26 March 2026  01:57:51 +0000 (0:00:12.112)       0:00:54.765 ******** \nPausing for 60 seconds\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***\nThursday 26 March 2026  01:58:51 +0000 (0:01:00.088)       0:01:54.854 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****\nThursday 26 March 2026  01:58:51 +0000 (0:00:00.074)       0:01:54.928 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  01:58:52 +0000 (0:00:00.567)       0:01:55.495 ******** \n=============================================================================== \nosism.services.squid : Wait for squid service to start ----------------- 60.09s\nosism.services.squid : Manage squid service ---------------------------- 36.63s\nosism.services.squid : Restart squid service --------------------------- 12.11s\nosism.services.squid : Install required packages ------------------------ 1.92s\nosism.services.squid : Create required directories ---------------------- 1.32s\nosism.services.squid : Copy squid configuration files ------------------- 1.15s\nosism.services.squid : Copy docker-compose.yml file --------------------- 0.96s\nosism.services.squid : Wait for an healthy squid service ---------------- 0.57s\nosism.services.squid : Remove osism_allow_list.conf configuration file --- 0.39s\nosism.services.squid : Include install tasks ---------------------------- 0.11s\nosism.services.squid : Register that squid service was restarted -------- 0.07s\n+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]\n++ semver 9.5.0 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release\n+ set -e\n+ NAMESPACE=kolla/release\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n++ semver 9.5.0 9.0.0\n+ [[ 1 -lt 0 ]]\n+ osism apply operator -u ubuntu -l testbed-nodes\n\u001b[32m2026-03-26 01:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b14432b8-054d-4f4f-bffe-fa6ab43829d4 (operator) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b14432b8-054d-4f4f-bffe-fa6ab43829d4 (operator) has been started and output is visible here.\u001b[0m\n\nPLAY [Make ssh pipelining working] *********************************************\n\nTASK [Gathering Facts] *********************************************************\nThursday 26 March 2026  01:59:08 +0000 (0:00:00.152)       0:00:00.152 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Do not require tty for all users] ****************************************\nThursday 26 March 2026  01:59:12 +0000 (0:00:03.324)       0:00:03.477 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply role operator] *****************************************************\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nThursday 26 March 2026  01:59:12 +0000 (0:00:00.753)       0:00:04.230 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nThursday 26 March 2026  01:59:13 +0000 (0:00:00.195)       0:00:04.425 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create operator group] **************************\nThursday 26 March 2026  01:59:13 +0000 (0:00:00.195)       0:00:04.621 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create user] ************************************\nThursday 26 March 2026  01:59:13 +0000 (0:00:00.624)       0:00:05.246 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nThursday 26 March 2026  01:59:14 +0000 (0:00:00.874)       0:00:06.120 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nThursday 26 March 2026  01:59:15 +0000 (0:00:01.247)       0:00:07.367 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nThursday 26 March 2026  01:59:17 +0000 (0:00:01.281)       0:00:08.648 ******** \n\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m\n\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m\n\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nThursday 26 March 2026  01:59:18 +0000 (0:00:01.301)       0:00:09.950 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nThursday 26 March 2026  01:59:18 +0000 (0:00:00.198)       0:00:10.148 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nThursday 26 March 2026  01:59:18 +0000 (0:00:00.187)       0:00:10.336 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nThursday 26 March 2026  01:59:19 +0000 (0:00:00.585)       0:00:10.921 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nThursday 26 March 2026  01:59:19 +0000 (0:00:00.184)       0:00:11.105 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nThursday 26 March 2026  01:59:20 +0000 (0:00:00.722)       0:00:11.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nThursday 26 March 2026  01:59:20 +0000 (0:00:00.177)       0:00:12.006 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nThursday 26 March 2026  01:59:20 +0000 (0:00:00.177)       0:00:12.183 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set password] ***********************************\nThursday 26 March 2026  01:59:20 +0000 (0:00:00.168)       0:00:12.352 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nThursday 26 March 2026  01:59:21 +0000 (0:00:00.650)       0:00:13.003 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  01:59:21 +0000 (0:00:00.269)       0:00:13.273 ******** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.32s\nosism.commons.operator : Set language variables in .bashrc configuration file --- 1.30s\nosism.commons.operator : Copy user sudoers file ------------------------- 1.28s\nosism.commons.operator : Add user to additional groups ------------------ 1.25s\nosism.commons.operator : Create user ------------------------------------ 0.87s\nDo not require tty for all users ---------------------------------------- 0.75s\nosism.commons.operator : Set ssh authorized keys ------------------------ 0.72s\nosism.commons.operator : Set password ----------------------------------- 0.65s\nosism.commons.operator : Create operator group -------------------------- 0.62s\nosism.commons.operator : Create .ssh directory -------------------------- 0.59s\nosism.commons.operator : Unset & lock password -------------------------- 0.27s\nosism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.20s\nosism.commons.operator : Gather variables for each operating system ----- 0.20s\nosism.commons.operator : Set operator_groups variable to default value --- 0.20s\nosism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.19s\nosism.commons.operator : Check number of SSH authorized keys ------------ 0.18s\nosism.commons.operator : Delete ssh authorized keys --------------------- 0.18s\nosism.commons.operator : Set authorized GitHub accounts ----------------- 0.18s\nosism.commons.operator : Delete authorized GitHub accounts -------------- 0.17s\n+ osism apply --environment custom facts\n\u001b[32m2026-03-26 01:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m\n\u001b[32m2026-03-26 01:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5fcfbccf-d31a-4b57-955a-42b16c057297 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 01:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5fcfbccf-d31a-4b57-955a-42b16c057297 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy custom network devices fact] ****************************************\n\nTASK [Create custom facts directory] *******************************************\nThursday 26 March 2026  01:59:38 +0000 (0:00:00.093)       0:00:00.093 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Copy fact file] **********************************************************\nThursday 26 March 2026  01:59:40 +0000 (0:00:01.381)       0:00:01.475 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Copy custom ceph devices facts] ******************************************\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nThursday 26 March 2026  01:59:41 +0000 (0:00:01.181)       0:00:02.657 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nThursday 26 March 2026  01:59:41 +0000 (0:00:00.110)       0:00:02.767 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nThursday 26 March 2026  01:59:41 +0000 (0:00:00.205)       0:00:02.973 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nThursday 26 March 2026  01:59:41 +0000 (0:00:00.238)       0:00:03.211 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nThursday 26 March 2026  01:59:41 +0000 (0:00:00.152)       0:00:03.363 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nThursday 26 March 2026  01:59:42 +0000 (0:00:00.418)       0:00:03.782 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nThursday 26 March 2026  01:59:42 +0000 (0:00:00.139)       0:00:03.921 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nThursday 26 March 2026  01:59:43 +0000 (0:00:01.063)       0:00:04.985 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nThursday 26 March 2026  01:59:43 +0000 (0:00:00.474)       0:00:05.459 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nThursday 26 March 2026  01:59:45 +0000 (0:00:01.094)       0:00:06.554 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Install required packages (RedHat)] **************************************\nThursday 26 March 2026  02:00:01 +0000 (0:00:16.562)       0:00:23.116 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (Debian)] **************************************\nThursday 26 March 2026  02:00:01 +0000 (0:00:00.105)       0:00:23.221 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Create custom facts directory] *******************************************\nThursday 26 March 2026  02:00:09 +0000 (0:00:07.791)       0:00:31.013 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Copy fact files] *********************************************************\nThursday 26 March 2026  02:00:10 +0000 (0:00:00.457)       0:00:31.470 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nThursday 26 March 2026  02:00:13 +0000 (0:00:03.462)       0:00:34.933 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:00:14 +0000 (0:00:01.500)       0:00:36.434 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:00:18 +0000 (0:00:03.653)       0:00:40.088 ******** \n=============================================================================== \nosism.commons.repository : Update package cache ------------------------ 16.56s\nInstall required packages (Debian) -------------------------------------- 7.79s\nGathers facts about hosts ----------------------------------------------- 3.65s\nCopy fact files --------------------------------------------------------- 3.46s\nosism.commons.repository : Force update of package cache ---------------- 1.50s\nCreate custom facts directory ------------------------------------------- 1.38s\nCopy fact file ---------------------------------------------------------- 1.18s\nosism.commons.repository : Copy ubuntu.sources file --------------------- 1.09s\nosism.commons.repository : Copy 99osism apt configuration --------------- 1.06s\nosism.commons.repository : Remove sources.list file --------------------- 0.47s\nCreate custom facts directory ------------------------------------------- 0.46s\nosism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.42s\nosism.commons.repository : Set repositories to default ------------------ 0.24s\nosism.commons.repository : Set repository_default fact to default value --- 0.21s\nosism.commons.repository : Include distribution specific repository tasks --- 0.15s\nosism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.14s\nosism.commons.repository : Gather variables for each operating system --- 0.11s\nInstall required packages (RedHat) -------------------------------------- 0.11s\n+ osism apply bootstrap\n\u001b[32m2026-03-26 02:00:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4680eed5-a920-43b6-870f-1050432d97e6 (bootstrap) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:00:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4680eed5-a920-43b6-870f-1050432d97e6 (bootstrap) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on state bootstrap] ************************************\n\nTASK [Group hosts based on state bootstrap] ************************************\nThursday 26 March 2026  02:00:36 +0000 (0:00:00.171)       0:00:00.171 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:00:36 +0000 (0:00:00.288)       0:00:00.459 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts (if using --limit)] ***************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:00:40 +0000 (0:00:03.598)       0:00:04.058 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 1] ********************************************\n\nTASK [osism.commons.hostname : Set hostname] ***********************************\nThursday 26 March 2026  02:00:40 +0000 (0:00:00.675)       0:00:04.734 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.hostname : Copy /etc/hostname] *****************************\nThursday 26 March 2026  02:00:41 +0000 (0:00:01.214)       0:00:05.949 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.hosts : Include type specific tasks] ***********************\nThursday 26 March 2026  02:00:43 +0000 (0:00:01.195)       0:00:07.145 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.hosts : Copy /etc/hosts file] ******************************\nThursday 26 March 2026  02:00:43 +0000 (0:00:00.304)       0:00:07.450 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.proxy : Include distribution specific tasks] ***************\nThursday 26 March 2026  02:00:45 +0000 (0:00:02.118)       0:00:09.568 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.proxy : Configure proxy parameters for apt] ****************\nThursday 26 March 2026  02:00:45 +0000 (0:00:00.305)       0:00:09.874 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.proxy : Set system wide settings in environment file] ******\nThursday 26 March 2026  02:00:46 +0000 (0:00:00.996)       0:00:10.871 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.proxy : Remove system wide settings in environment file] ***\nThursday 26 March 2026  02:00:47 +0000 (0:00:00.561)       0:00:11.433 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nThursday 26 March 2026  02:00:47 +0000 (0:00:00.457)       0:00:11.890 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nThursday 26 March 2026  02:00:48 +0000 (0:00:00.243)       0:00:12.134 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nThursday 26 March 2026  02:00:48 +0000 (0:00:00.303)       0:00:12.437 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nThursday 26 March 2026  02:00:48 +0000 (0:00:00.326)       0:00:12.764 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nThursday 26 March 2026  02:00:50 +0000 (0:00:01.380)       0:00:14.144 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nThursday 26 March 2026  02:00:50 +0000 (0:00:00.353)       0:00:14.497 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nThursday 26 March 2026  02:00:51 +0000 (0:00:00.559)       0:00:15.057 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nThursday 26 March 2026  02:00:51 +0000 (0:00:00.281)       0:00:15.339 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nThursday 26 March 2026  02:00:51 +0000 (0:00:00.557)       0:00:15.896 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nThursday 26 March 2026  02:00:53 +0000 (0:00:01.134)       0:00:17.030 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nThursday 26 March 2026  02:00:54 +0000 (0:00:01.025)       0:00:18.056 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nThursday 26 March 2026  02:00:54 +0000 (0:00:00.329)       0:00:18.385 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nThursday 26 March 2026  02:00:55 +0000 (0:00:01.274)       0:00:19.660 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nThursday 26 March 2026  02:00:55 +0000 (0:00:00.243)       0:00:19.903 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nThursday 26 March 2026  02:00:56 +0000 (0:00:00.225)       0:00:20.129 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nThursday 26 March 2026  02:00:56 +0000 (0:00:00.274)       0:00:20.403 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nThursday 26 March 2026  02:00:56 +0000 (0:00:00.345)       0:00:20.749 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nThursday 26 March 2026  02:00:57 +0000 (0:00:00.521)       0:00:21.270 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nThursday 26 March 2026  02:00:57 +0000 (0:00:00.289)       0:00:21.560 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nThursday 26 March 2026  02:00:58 +0000 (0:00:01.103)       0:00:22.663 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nThursday 26 March 2026  02:00:59 +0000 (0:00:00.571)       0:00:23.235 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nThursday 26 March 2026  02:01:00 +0000 (0:00:01.168)       0:00:24.404 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Gather variables for each operating system] *****\nThursday 26 March 2026  02:01:16 +0000 (0:00:16.032)       0:00:40.436 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****\nThursday 26 March 2026  02:01:16 +0000 (0:00:00.214)       0:00:40.651 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***\nThursday 26 March 2026  02:01:16 +0000 (0:00:00.239)       0:00:40.890 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include distribution specific install tasks] ****\nThursday 26 March 2026  02:01:17 +0000 (0:00:00.239)       0:00:41.130 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Install rsyslog package] ************************\nThursday 26 March 2026  02:01:17 +0000 (0:00:00.333)       0:00:41.463 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********\nThursday 26 March 2026  02:01:19 +0000 (0:00:01.723)       0:00:43.187 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Manage rsyslog service] *************************\nThursday 26 March 2026  02:01:20 +0000 (0:00:01.058)       0:00:44.246 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include fluentd tasks] **************************\nThursday 26 March 2026  02:01:21 +0000 (0:00:00.824)       0:00:45.070 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***\nThursday 26 March 2026  02:01:21 +0000 (0:00:00.347)       0:00:45.418 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include additional log server tasks] ************\nThursday 26 March 2026  02:01:22 +0000 (0:00:01.002)       0:00:46.420 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include logrotate tasks] ************************\nThursday 26 March 2026  02:01:22 +0000 (0:00:00.255)       0:00:46.675 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Ensure logrotate package is installed] **********\nThursday 26 March 2026  02:01:22 +0000 (0:00:00.330)       0:00:47.006 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************\nThursday 26 March 2026  02:01:24 +0000 (0:00:01.779)       0:00:48.786 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.systohc : Install util-linux-extra package] ****************\nThursday 26 March 2026  02:01:25 +0000 (0:00:01.139)       0:00:49.925 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.systohc : Sync hardware clock] *****************************\nThursday 26 March 2026  02:01:39 +0000 (0:00:13.497)       0:01:03.423 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.configfs : Start sys-kernel-config mount] ******************\nThursday 26 March 2026  02:01:40 +0000 (0:00:01.248)       0:01:04.671 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Gather variables for each operating system] *****\nThursday 26 March 2026  02:01:41 +0000 (0:00:00.945)       0:01:05.617 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***\nThursday 26 March 2026  02:01:41 +0000 (0:00:00.272)       0:01:05.889 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Include distribution specific package tasks] ****\nThursday 26 March 2026  02:01:42 +0000 (0:00:00.253)       0:01:06.143 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.packages : Install needrestart package] ********************\nThursday 26 March 2026  02:01:42 +0000 (0:00:00.316)       0:01:06.459 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.packages : Set needrestart mode] ***************************\nThursday 26 March 2026  02:01:44 +0000 (0:00:01.604)       0:01:08.063 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***\nThursday 26 March 2026  02:01:44 +0000 (0:00:00.565)       0:01:08.628 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Update package cache] ***************************\nThursday 26 March 2026  02:01:44 +0000 (0:00:00.241)       0:01:08.870 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.packages : Download upgrade packages] **********************\nThursday 26 March 2026  02:01:45 +0000 (0:00:01.143)       0:01:10.013 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.packages : Upgrade packages] *******************************\nThursday 26 March 2026  02:01:47 +0000 (0:00:01.661)       0:01:11.674 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Download required packages] *********************\nThursday 26 March 2026  02:01:50 +0000 (0:00:03.133)       0:01:14.808 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Install required packages] **********************\nThursday 26 March 2026  02:02:35 +0000 (0:00:44.819)       0:01:59.627 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Remove useless packages from the cache] *********\nThursday 26 March 2026  02:03:50 +0000 (0:01:14.819)       0:03:14.446 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Remove dependencies that are no longer required] ***\nThursday 26 March 2026  02:03:52 +0000 (0:00:01.685)       0:03:16.132 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sysctl : Include sysctl tasks] *****************************\nThursday 26 March 2026  02:04:07 +0000 (0:00:15.146)       0:03:31.279 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********\nThursday 26 March 2026  02:04:07 +0000 (0:00:00.462)       0:03:31.741 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************\nThursday 26 March 2026  02:04:08 +0000 (0:00:00.829)       0:03:32.571 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************\nThursday 26 March 2026  02:04:13 +0000 (0:00:04.830)       0:03:37.401 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************\nThursday 26 March 2026  02:04:13 +0000 (0:00:00.583)       0:03:37.985 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on network] *****************\nThursday 26 March 2026  02:04:14 +0000 (0:00:00.524)       0:03:38.509 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************\nThursday 26 March 2026  02:04:15 +0000 (0:00:00.605)       0:03:39.115 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\nTASK [osism.commons.limits : Include limits tasks] *****************************\nThursday 26 March 2026  02:04:15 +0000 (0:00:00.611)       0:03:39.727 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Populate service facts] *************************\nThursday 26 March 2026  02:04:16 +0000 (0:00:00.365)       0:03:40.092 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.services : Check services] *********************************\nThursday 26 March 2026  02:04:22 +0000 (0:00:06.599)       0:03:46.692 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Start/enable required services] *****************\nThursday 26 March 2026  02:04:23 +0000 (0:00:00.338)       0:03:47.031 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m\n\nTASK [osism.commons.motd : Include distribution specific configure tasks] ******\nThursday 26 March 2026  02:04:24 +0000 (0:00:01.163)       0:03:48.195 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.motd : Remove update-motd package] *************************\nThursday 26 March 2026  02:04:24 +0000 (0:00:00.448)       0:03:48.643 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************\nThursday 26 March 2026  02:04:26 +0000 (0:00:01.429)       0:03:50.072 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Disable the dynamic motd-news service] **************\nThursday 26 March 2026  02:04:26 +0000 (0:00:00.660)       0:03:50.732 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********\nThursday 26 March 2026  02:04:27 +0000 (0:00:00.640)       0:03:51.373 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Remove pam_motd.so rule] ****************************\nThursday 26 March 2026  02:04:27 +0000 (0:00:00.601)       0:03:51.975 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489204.3152304, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489229.2050931, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489232.927024, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489223.8100085, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489221.6920967, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489235.4309287, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489217.9835355, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [osism.commons.motd : Copy motd file] *************************************\nThursday 26 March 2026  02:04:28 +0000 (0:00:00.949)       0:03:52.925 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue file] ************************************\nThursday 26 March 2026  02:04:29 +0000 (0:00:01.067)       0:03:53.992 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue.net file] ********************************\nThursday 26 March 2026  02:04:31 +0000 (0:00:01.206)       0:03:55.199 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to print the motd] ********************\nThursday 26 March 2026  02:04:32 +0000 (0:00:01.160)       0:03:56.359 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to not print the motd] ****************\nThursday 26 March 2026  02:04:32 +0000 (0:00:00.362)       0:03:56.722 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.rng : Include distribution specific install tasks] ********\nThursday 26 March 2026  02:04:33 +0000 (0:00:00.797)       0:03:57.519 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rng : Install rng package] ********************************\nThursday 26 March 2026  02:04:34 +0000 (0:00:00.518)       0:03:58.037 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rng : Remove haveged package] *****************************\nThursday 26 March 2026  02:04:42 +0000 (0:00:08.264)       0:04:06.302 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rng : Manage rng service] *********************************\nThursday 26 March 2026  02:04:43 +0000 (0:00:01.293)       0:04:07.595 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Gather variables for each operating system] ******\nThursday 26 March 2026  02:04:44 +0000 (0:00:01.132)       0:04:08.728 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***\nThursday 26 March 2026  02:04:45 +0000 (0:00:00.334)       0:04:09.063 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***\nThursday 26 March 2026  02:04:45 +0000 (0:00:00.329)       0:04:09.392 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Populate service facts] **************************\nThursday 26 March 2026  02:04:45 +0000 (0:00:00.344)       0:04:09.737 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Include distribution specific timer tasks] *******\nThursday 26 March 2026  02:04:51 +0000 (0:00:05.484)       0:04:15.221 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Disable apt-daily timers] ************************\nThursday 26 March 2026  02:04:51 +0000 (0:00:00.456)       0:04:15.678 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include service tasks] ***************************\nThursday 26 March 2026  02:04:52 +0000 (0:00:00.432)       0:04:16.111 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup services] ********************************\nThursday 26 March 2026  02:04:52 +0000 (0:00:00.463)       0:04:16.574 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include packages tasks] **************************\nThursday 26 March 2026  02:04:52 +0000 (0:00:00.391)       0:04:16.965 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup installed packages] **********************\nThursday 26 March 2026  02:04:53 +0000 (0:00:00.463)       0:04:17.429 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloudinit package] ************************\nThursday 26 March 2026  02:05:28 +0000 (0:00:35.079)       0:04:52.509 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********\nThursday 26 March 2026  02:05:36 +0000 (0:00:08.100)       0:05:00.609 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove useless packages from the cache] **********\nThursday 26 March 2026  02:05:44 +0000 (0:00:08.292)       0:05:08.902 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***\nThursday 26 March 2026  02:05:46 +0000 (0:00:01.785)       0:05:10.688 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Include cloudinit tasks] *************************\nThursday 26 March 2026  02:05:52 +0000 (0:00:06.100)       0:05:16.788 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******\nThursday 26 March 2026  02:05:53 +0000 (0:00:00.464)       0:05:17.253 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Install tzdata package] *************************\nThursday 26 March 2026  02:05:53 +0000 (0:00:00.750)       0:05:18.003 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.timezone : Set timezone to UTC] ****************************\nThursday 26 March 2026  02:05:55 +0000 (0:00:01.731)       0:05:19.735 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.timezone : Create /etc/adjtime file] ***********************\nThursday 26 March 2026  02:05:56 +0000 (0:00:00.814)       0:05:20.550 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************\nThursday 26 March 2026  02:05:56 +0000 (0:00:00.365)       0:05:20.915 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Gather variables for each operating system] ******\nThursday 26 March 2026  02:05:57 +0000 (0:00:00.500)       0:05:21.416 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_version variable to default value] ****\nThursday 26 March 2026  02:05:57 +0000 (0:00:00.403)       0:05:21.819 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_cli_version variable to default value] ***\nThursday 26 March 2026  02:05:58 +0000 (0:00:00.364)       0:05:22.183 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Print used docker version] ***********************\nThursday 26 March 2026  02:05:58 +0000 (0:00:00.356)       0:05:22.539 ******** \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Print used docker cli version] *******************\nThursday 26 March 2026  02:05:58 +0000 (0:00:00.331)       0:05:22.871 ******** \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Include block storage tasks] *********************\nThursday 26 March 2026  02:05:59 +0000 (0:00:00.349)       0:05:23.221 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include zram storage tasks] **********************\nThursday 26 March 2026  02:05:59 +0000 (0:00:00.327)       0:05:23.548 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include docker install tasks] ********************\nThursday 26 March 2026  02:05:59 +0000 (0:00:00.306)       0:05:23.854 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Remove old architecture-dependent repository] ****\nThursday 26 March 2026  02:06:00 +0000 (0:00:00.433)       0:05:24.288 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Gather package facts] ****************************\nThursday 26 March 2026  02:06:01 +0000 (0:00:01.019)       0:05:25.308 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Check whether packages are installed that should not be installed] ***\nThursday 26 March 2026  02:06:04 +0000 (0:00:03.173)       0:05:28.482 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install apt-transport-https package] *************\nThursday 26 March 2026  02:06:05 +0000 (0:00:00.724)       0:05:29.207 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Add repository gpg key] **************************\nThursday 26 March 2026  02:06:11 +0000 (0:00:06.691)       0:05:35.898 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Add repository] **********************************\nThursday 26 March 2026  02:06:13 +0000 (0:00:01.151)       0:05:37.050 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Update package cache] ****************************\nThursday 26 March 2026  02:06:21 +0000 (0:00:08.358)       0:05:45.408 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Pin docker package version] **********************\nThursday 26 March 2026  02:06:24 +0000 (0:00:03.275)       0:05:48.684 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Pin docker-cli package version] ******************\nThursday 26 March 2026  02:06:26 +0000 (0:00:01.393)       0:05:50.077 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Unlock containerd package] ***********************\nThursday 26 March 2026  02:06:27 +0000 (0:00:01.803)       0:05:51.881 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install containerd package] **********************\nThursday 26 March 2026  02:06:28 +0000 (0:00:00.661)       0:05:52.542 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Lock containerd package] *************************\nThursday 26 March 2026  02:06:37 +0000 (0:00:08.552)       0:06:01.095 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker-cli package] **********************\nThursday 26 March 2026  02:06:38 +0000 (0:00:01.018)       0:06:02.114 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker package] **************************\nThursday 26 March 2026  02:06:46 +0000 (0:00:08.642)       0:06:10.756 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Unblock installation of python docker packages] ***\nThursday 26 March 2026  02:06:57 +0000 (0:00:10.659)       0:06:21.415 ******** \n\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package] ******************\nThursday 26 March 2026  02:06:58 +0000 (0:00:01.231)       0:06:22.647 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package from Debian Sid] ***\nThursday 26 March 2026  02:06:59 +0000 (0:00:00.659)       0:06:23.306 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***\nThursday 26 March 2026  02:07:03 +0000 (0:00:03.719)       0:06:27.026 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***\nThursday 26 March 2026  02:07:03 +0000 (0:00:00.547)       0:06:27.574 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***\nThursday 26 March 2026  02:07:04 +0000 (0:00:00.853)       0:06:28.427 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***\nThursday 26 March 2026  02:07:04 +0000 (0:00:00.579)       0:06:29.006 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install packages required by docker login] *******\nThursday 26 March 2026  02:07:05 +0000 (0:00:00.498)       0:06:29.504 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Ensure that some packages are not installed] *****\nThursday 26 March 2026  02:07:06 +0000 (0:00:00.579)       0:06:30.084 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include config tasks] ****************************\nThursday 26 March 2026  02:07:08 +0000 (0:00:02.001)       0:06:32.086 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create plugins directory] ************************\nThursday 26 March 2026  02:07:09 +0000 (0:00:01.015)       0:06:33.101 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Create systemd overlay directory] ****************\nThursday 26 March 2026  02:07:09 +0000 (0:00:00.889)       0:06:33.991 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy systemd overlay file] ***********************\nThursday 26 March 2026  02:07:10 +0000 (0:00:00.846)       0:06:34.838 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***\nThursday 26 March 2026  02:07:12 +0000 (0:00:01.747)       0:06:36.585 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy limits configuration file] ******************\nThursday 26 March 2026  02:07:14 +0000 (0:00:01.485)       0:06:38.070 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy daemon.json configuration file] *************\nThursday 26 March 2026  02:07:15 +0000 (0:00:01.332)       0:06:39.403 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include service tasks] ***************************\nThursday 26 March 2026  02:07:16 +0000 (0:00:01.554)       0:06:40.958 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon] ***************************\nThursday 26 March 2026  02:07:18 +0000 (0:00:01.118)       0:06:42.076 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage service] **********************************\nThursday 26 March 2026  02:07:19 +0000 (0:00:01.324)       0:06:43.400 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Manage docker socket service] ********************\nThursday 26 March 2026  02:07:21 +0000 (0:00:02.024)       0:06:45.425 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage containerd service] ***********************\nThursday 26 March 2026  02:07:22 +0000 (0:00:01.150)       0:06:46.576 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include bootstrap tasks] *************************\nThursday 26 March 2026  02:07:23 +0000 (0:00:01.315)       0:06:47.891 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:24 +0000 (0:00:01.010)       0:06:48.902 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:24 +0000 (0:00:00.043)       0:06:48.946 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:24 +0000 (0:00:00.049)       0:06:48.995 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:25 +0000 (0:00:00.043)       0:06:49.039 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:25 +0000 (0:00:00.048)       0:06:49.088 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:25 +0000 (0:00:00.048)       0:06:49.136 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nThursday 26 March 2026  02:07:25 +0000 (0:00:00.040)       0:06:49.177 ******** \n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nThursday 26 March 2026  02:07:25 +0000 (0:00:00.042)       0:06:49.220 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************\nThursday 26 March 2026  02:07:26 +0000 (0:00:01.242)       0:06:50.463 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********\nThursday 26 March 2026  02:07:28 +0000 (0:00:01.569)       0:06:52.032 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Restart docker service] ***************\nThursday 26 March 2026  02:07:29 +0000 (0:00:01.262)       0:06:53.294 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****\nThursday 26 March 2026  02:07:31 +0000 (0:00:02.423)       0:06:55.717 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Add user to docker group] ************************\nThursday 26 March 2026  02:07:31 +0000 (0:00:00.106)       0:06:55.824 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Log into private registry and force re-authorization] ***\nThursday 26 March 2026  02:07:32 +0000 (0:00:01.117)       0:06:56.942 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include facts tasks] *****************************\nThursday 26 March 2026  02:07:33 +0000 (0:00:00.632)       0:06:57.574 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create facts directory] **************************\nThursday 26 March 2026  02:07:34 +0000 (0:00:01.197)       0:06:58.772 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy docker fact files] **************************\nThursday 26 March 2026  02:07:35 +0000 (0:00:00.853)       0:06:59.625 ******** \n\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m\n\nTASK [osism.commons.docker_compose : This install type is not supported] *******\nThursday 26 March 2026  02:07:38 +0000 (0:00:02.520)       0:07:02.146 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Include distribution specific install tasks] ***\nThursday 26 March 2026  02:07:38 +0000 (0:00:00.720)       0:07:02.866 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***\nThursday 26 March 2026  02:07:39 +0000 (0:00:00.835)       0:07:03.701 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******\nThursday 26 March 2026  02:07:40 +0000 (0:00:00.850)       0:07:04.552 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose binary] *************\nThursday 26 March 2026  02:07:41 +0000 (0:00:01.104)       0:07:05.656 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Uninstall docker-compose package] *********\nThursday 26 March 2026  02:07:42 +0000 (0:00:00.544)       0:07:06.201 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose script] ***************\nThursday 26 March 2026  02:07:43 +0000 (0:00:01.498)       0:07:07.700 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****\nThursday 26 March 2026  02:07:44 +0000 (0:00:00.616)       0:07:08.316 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********\nThursday 26 March 2026  02:07:52 +0000 (0:00:07.978)       0:07:16.295 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Enable osism.target] **********************\nThursday 26 March 2026  02:07:53 +0000 (0:00:01.620)       0:07:17.916 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****\nThursday 26 March 2026  02:07:55 +0000 (0:00:01.641)       0:07:19.557 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nThursday 26 March 2026  02:07:57 +0000 (0:00:01.627)       0:07:21.185 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nThursday 26 March 2026  02:07:58 +0000 (0:00:00.872)       0:07:22.057 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check minimum and maximum number of servers] *****\nThursday 26 March 2026  02:07:59 +0000 (0:00:01.129)       0:07:23.186 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Gather variables for each operating system] ******\nThursday 26 March 2026  02:07:59 +0000 (0:00:00.559)       0:07:23.746 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***\nThursday 26 March 2026  02:08:00 +0000 (0:00:00.580)       0:07:24.326 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_key_file variable to default value] ***\nThursday 26 March 2026  02:08:01 +0000 (0:00:00.796)       0:07:25.123 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Populate service facts] **************************\nThursday 26 March 2026  02:08:01 +0000 (0:00:00.570)       0:07:25.693 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.chrony : Manage timesyncd service] ************************\nThursday 26 March 2026  02:08:07 +0000 (0:00:05.626)       0:07:31.319 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Include distribution specific install tasks] *****\nThursday 26 March 2026  02:08:07 +0000 (0:00:00.601)       0:07:31.921 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.chrony : Install package] *********************************\nThursday 26 March 2026  02:08:09 +0000 (0:00:01.100)       0:07:33.021 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.chrony : Manage chrony service] ***************************\nThursday 26 March 2026  02:08:10 +0000 (0:00:01.951)       0:07:34.973 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check if configuration file exists] **************\nThursday 26 March 2026  02:08:12 +0000 (0:00:01.256)       0:07:36.229 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Copy configuration file] *************************\nThursday 26 March 2026  02:08:13 +0000 (0:00:00.875)       0:07:37.105 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\nTASK [osism.services.lldpd : Include distribution specific install tasks] ******\nThursday 26 March 2026  02:08:15 +0000 (0:00:02.118)       0:07:39.224 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.lldpd : Install lldpd package] ****************************\nThursday 26 March 2026  02:08:16 +0000 (0:00:00.919)       0:07:40.143 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.lldpd : Manage lldpd service] *****************************\nThursday 26 March 2026  02:08:25 +0000 (0:00:09.433)       0:07:49.577 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********\nThursday 26 March 2026  02:08:27 +0000 (0:00:01.978)       0:07:51.555 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************\nThursday 26 March 2026  02:08:28 +0000 (0:00:01.217)       0:07:52.773 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap role part 2] *********************************************\n\nTASK [Include hardening role] **************************************************\nThursday 26 March 2026  02:08:30 +0000 (0:00:01.322)       0:07:54.096 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 3] ********************************************\n\nTASK [osism.services.journald : Copy configuration file] ***********************\nThursday 26 March 2026  02:08:30 +0000 (0:00:00.768)       0:07:54.865 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.journald : Manage journald service] ***********************\nThursday 26 March 2026  02:08:32 +0000 (0:00:01.260)       0:07:56.126 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include auditd role] *****************************************************\nThursday 26 March 2026  02:08:33 +0000 (0:00:01.433)       0:07:57.559 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include smartd role] *****************************************************\nThursday 26 March 2026  02:08:34 +0000 (0:00:00.536)       0:07:58.095 ******** \n\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Include distribution specific install tasks] *****\nThursday 26 March 2026  02:08:35 +0000 (0:00:01.101)       0:07:59.197 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Install smartmontools package] *******************\nThursday 26 March 2026  02:08:36 +0000 (0:00:00.910)       0:08:00.108 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.smartd : Create /var/log/smartd directory] ****************\nThursday 26 March 2026  02:08:44 +0000 (0:00:08.733)       0:08:08.841 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Copy smartmontools configuration file] ***********\nThursday 26 March 2026  02:08:45 +0000 (0:00:00.879)       0:08:09.721 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Manage smartd service] ***************************\nThursday 26 March 2026  02:08:47 +0000 (0:00:01.345)       0:08:11.067 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.journald : Restart journald service] ***********\nThursday 26 March 2026  02:08:49 +0000 (0:00:02.143)       0:08:13.211 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************\nThursday 26 March 2026  02:08:50 +0000 (0:00:01.251)       0:08:14.462 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Set state bootstrap] *****************************************************\n\nTASK [Set osism.bootstrap.status fact] *****************************************\nThursday 26 March 2026  02:08:51 +0000 (0:00:01.174)       0:08:15.636 ******** \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nThursday 26 March 2026  02:08:52 +0000 (0:00:00.861)       0:08:16.497 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nThursday 26 March 2026  02:08:53 +0000 (0:00:01.080)       0:08:17.578 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Set osism.bootstrap.timestamp fact] **************************************\nThursday 26 March 2026  02:08:54 +0000 (0:00:01.166)       0:08:18.745 ******** \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nThursday 26 March 2026  02:08:55 +0000 (0:00:01.081)       0:08:19.826 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nThursday 26 March 2026  02:08:56 +0000 (0:00:00.897)       0:08:20.723 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:08:57 +0000 (0:00:01.223)       0:08:21.947 ******** \n=============================================================================== \nosism.commons.packages : Install required packages --------------------- 74.82s\nosism.commons.packages : Download required packages -------------------- 44.82s\nosism.commons.cleanup : Cleanup installed packages --------------------- 35.08s\nosism.commons.repository : Update package cache ------------------------ 16.03s\nosism.commons.packages : Remove dependencies that are no longer required -- 15.15s\nosism.commons.systohc : Install util-linux-extra package --------------- 13.50s\nosism.services.docker : Install docker package ------------------------- 10.66s\nosism.services.lldpd : Install lldpd package ---------------------------- 9.43s\nosism.services.smartd : Install smartmontools package ------------------- 8.73s\nosism.services.docker : Install docker-cli package ---------------------- 8.64s\nosism.services.docker : Install containerd package ---------------------- 8.55s\nosism.services.docker : Add repository ---------------------------------- 8.36s\nosism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.29s\nosism.services.rng : Install rng package -------------------------------- 8.26s\nosism.commons.cleanup : Remove cloudinit package ------------------------ 8.10s\nosism.commons.docker_compose : Install docker-compose-plugin package ---- 7.98s\nosism.services.docker : Install apt-transport-https package ------------- 6.69s\nosism.commons.services : Populate service facts ------------------------- 6.60s\nosism.commons.cleanup : Remove dependencies that are no longer required --- 6.10s\nosism.services.chrony : Populate service facts -------------------------- 5.63s\n+ osism apply fail2ban\n\u001b[32m2026-03-26 02:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fcfdf61c-4c97-405f-bf69-c9730546e312 (fail2ban) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fcfdf61c-4c97-405f-bf69-c9730546e312 (fail2ban) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role fail2ban] *****************************************************\n\nTASK [osism.services.fail2ban : Include distribution specific install tasks] ***\nThursday 26 March 2026  02:09:17 +0000 (0:00:00.287)       0:00:00.287 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.fail2ban : Install fail2ban package] **********************\nThursday 26 March 2026  02:09:18 +0000 (0:00:01.216)       0:00:01.504 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.fail2ban : Copy configuration files] **********************\nThursday 26 March 2026  02:09:29 +0000 (0:00:11.356)       0:00:12.860 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Manage fail2ban service] ***********************\nThursday 26 March 2026  02:09:31 +0000 (0:00:01.453)       0:00:14.313 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Reload fail2ban configuration] *****************\nThursday 26 March 2026  02:09:32 +0000 (0:00:01.479)       0:00:15.793 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:09:34 +0000 (0:00:01.671)       0:00:17.465 ******** \n=============================================================================== \nosism.services.fail2ban : Install fail2ban package --------------------- 11.36s\nosism.services.fail2ban : Reload fail2ban configuration ----------------- 1.67s\nosism.services.fail2ban : Manage fail2ban service ----------------------- 1.48s\nosism.services.fail2ban : Copy configuration files ---------------------- 1.45s\nosism.services.fail2ban : Include distribution specific install tasks --- 1.22s\n+ [[ -e /etc/redhat-release ]]\n+ osism apply network\n\u001b[32m2026-03-26 02:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d321b30-b3f8-4748-8c86-c02309be4909 (network) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d321b30-b3f8-4748-8c86-c02309be4909 (network) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role network] ******************************************************\n\nTASK [osism.commons.network : Gather variables for each operating system] ******\nThursday 26 March 2026  02:09:52 +0000 (0:00:00.294)       0:00:00.294 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include type specific tasks] *********************\nThursday 26 March 2026  02:09:52 +0000 (0:00:00.779)       0:00:01.073 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Install required packages] ***********************\nThursday 26 March 2026  02:09:54 +0000 (0:00:01.368)       0:00:02.442 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.network : Remove ifupdown package] *************************\nThursday 26 March 2026  02:09:56 +0000 (0:00:02.009)       0:00:04.452 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Create required directories] *********************\nThursday 26 March 2026  02:09:58 +0000 (0:00:01.785)       0:00:06.237 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m\n\nTASK [osism.commons.network : Prepare netplan configuration template] **********\nThursday 26 March 2026  02:09:59 +0000 (0:00:00.991)       0:00:07.229 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Copy netplan configuration] **********************\nThursday 26 March 2026  02:10:03 +0000 (0:00:03.990)       0:00:11.220 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove netplan configuration template] ***********\nThursday 26 March 2026  02:10:04 +0000 (0:00:01.673)       0:00:12.894 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Check if path for interface file exists] *********\nThursday 26 March 2026  02:10:06 +0000 (0:00:01.953)       0:00:14.847 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy interfaces file] ****************************\nThursday 26 March 2026  02:10:07 +0000 (0:00:01.259)       0:00:16.107 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Install package networkd-dispatcher] *************\nThursday 26 March 2026  02:10:08 +0000 (0:00:00.764)       0:00:16.871 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy dispatcher scripts] *************************\nThursday 26 March 2026  02:10:10 +0000 (0:00:02.186)       0:00:19.058 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m\n\nTASK [osism.commons.network : Manage service networkd-dispatcher] **************\nThursday 26 March 2026  02:10:11 +0000 (0:00:01.044)       0:00:20.102 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include cleanup tasks] ***************************\nThursday 26 March 2026  02:10:13 +0000 (0:00:01.758)       0:00:21.860 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nThursday 26 March 2026  02:10:15 +0000 (0:00:01.358)       0:00:23.218 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Set network_configured_files fact] ***************\nThursday 26 March 2026  02:10:16 +0000 (0:00:01.181)       0:00:24.400 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nThursday 26 March 2026  02:10:16 +0000 (0:00:00.713)       0:00:25.113 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\nTASK [osism.commons.network : Include dummy interfaces] ************************\nThursday 26 March 2026  02:10:18 +0000 (0:00:01.255)       0:00:26.368 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include vxlan interfaces] ************************\nThursday 26 March 2026  02:10:18 +0000 (0:00:00.679)       0:00:27.048 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-manager, testbed-node-3, testbed-node-2, testbed-node-4, testbed-node-1, testbed-node-0, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd netdev files] ************\nThursday 26 March 2026  02:10:25 +0000 (0:00:06.448)       0:00:33.497 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd network files] ***********\nThursday 26 March 2026  02:10:32 +0000 (0:00:06.877)       0:00:40.374 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Include networkd cleanup tasks] ******************\nThursday 26 March 2026  02:10:38 +0000 (0:00:06.228)       0:00:46.602 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nThursday 26 March 2026  02:10:39 +0000 (0:00:01.442)       0:00:48.045 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nThursday 26 March 2026  02:10:41 +0000 (0:00:01.441)       0:00:49.487 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************\nThursday 26 March 2026  02:10:44 +0000 (0:00:02.662)       0:00:52.150 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********\nThursday 26 March 2026  02:10:44 +0000 (0:00:00.724)       0:00:52.874 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:10:45 +0000 (0:00:00.786)       0:00:53.661 ******** \n=============================================================================== \nosism.commons.network : Create systemd networkd netdev files ------------ 6.88s\nosism.commons.network : Include vxlan interfaces ------------------------ 6.45s\nosism.commons.network : Create systemd networkd network files ----------- 6.23s\nosism.commons.network : Prepare netplan configuration template ---------- 3.99s\nosism.commons.network : Remove unused configuration files --------------- 2.66s\nosism.commons.network : Install package networkd-dispatcher ------------- 2.19s\nosism.commons.network : Install required packages ----------------------- 2.01s\nosism.commons.network : Remove netplan configuration template ----------- 1.95s\nosism.commons.network : Remove ifupdown package ------------------------- 1.79s\nosism.commons.network : Manage service networkd-dispatcher -------------- 1.76s\nosism.commons.network : Copy netplan configuration ---------------------- 1.67s\nosism.commons.network : Include networkd cleanup tasks ------------------ 1.44s\nosism.commons.network : List existing configuration files --------------- 1.44s\nosism.commons.network : Include type specific tasks --------------------- 1.37s\nosism.commons.network : Include cleanup tasks --------------------------- 1.36s\nosism.commons.network : Check if path for interface file exists --------- 1.26s\nosism.commons.network : Remove unused configuration files --------------- 1.26s\nosism.commons.network : List existing configuration files --------------- 1.18s\nosism.commons.network : Copy dispatcher scripts ------------------------- 1.04s\nosism.commons.network : Create required directories --------------------- 0.99s\n+ osism apply wireguard\n\u001b[32m2026-03-26 02:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d2213621-6efb-4be0-b77b-16b9f0ccce98 (wireguard) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d2213621-6efb-4be0-b77b-16b9f0ccce98 (wireguard) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role wireguard] ****************************************************\n\nTASK [osism.services.wireguard : Install iptables package] *********************\nThursday 26 March 2026  02:11:03 +0000 (0:00:00.253)       0:00:00.253 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Install wireguard package] ********************\nThursday 26 March 2026  02:11:05 +0000 (0:00:01.799)       0:00:02.053 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create public and private key - server] *******\nThursday 26 March 2026  02:11:12 +0000 (0:00:07.246)       0:00:09.300 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create preshared key] *************************\nThursday 26 March 2026  02:11:12 +0000 (0:00:00.571)       0:00:09.872 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get preshared key] ****************************\nThursday 26 March 2026  02:11:13 +0000 (0:00:00.497)       0:00:10.370 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get public key - server] **********************\nThursday 26 March 2026  02:11:14 +0000 (0:00:00.790)       0:00:11.160 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get private key - server] *********************\nThursday 26 March 2026  02:11:14 +0000 (0:00:00.423)       0:00:11.583 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy wg0.conf configuration file] *************\nThursday 26 March 2026  02:11:15 +0000 (0:00:00.423)       0:00:12.007 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy client configuration files] **************\nThursday 26 March 2026  02:11:16 +0000 (0:00:01.252)       0:00:13.259 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********\nThursday 26 March 2026  02:11:17 +0000 (0:00:00.946)       0:00:14.206 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************\nThursday 26 March 2026  02:11:19 +0000 (0:00:01.789)       0:00:15.996 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:11:20 +0000 (0:00:00.949)       0:00:16.946 ******** \n=============================================================================== \nosism.services.wireguard : Install wireguard package -------------------- 7.25s\nosism.services.wireguard : Install iptables package --------------------- 1.80s\nosism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.79s\nosism.services.wireguard : Copy wg0.conf configuration file ------------- 1.25s\nosism.services.wireguard : Restart wg0 service -------------------------- 0.95s\nosism.services.wireguard : Copy client configuration files -------------- 0.95s\nosism.services.wireguard : Get preshared key ---------------------------- 0.79s\nosism.services.wireguard : Create public and private key - server ------- 0.57s\nosism.services.wireguard : Create preshared key ------------------------- 0.50s\nosism.services.wireguard : Get private key - server --------------------- 0.42s\nosism.services.wireguard : Get public key - server ---------------------- 0.42s\n+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0\r100    14  100    14    0     0    168      0 --:--:-- --:--:-- --:--:--   170\n+ osism apply --environment custom workarounds\n\u001b[32m2026-03-26 02:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m\n\u001b[32m2026-03-26 02:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 02ff3201-de62-47b0-bff2-96101897c72a (workarounds) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 02ff3201-de62-47b0-bff2-96101897c72a (workarounds) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on virtualization_role] ********************************\nThursday 26 March 2026  02:11:37 +0000 (0:00:00.131)       0:00:00.131 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m\n\nPLAY [Apply netplan configuration on the manager node] *************************\n\nTASK [Apply netplan configuration] *********************************************\nThursday 26 March 2026  02:11:38 +0000 (0:00:00.840)       0:00:00.972 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply netplan configuration on all other nodes] **************************\n\nTASK [Apply netplan configuration] *********************************************\nThursday 26 March 2026  02:11:41 +0000 (0:00:02.657)       0:00:03.629 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Add custom CA certificates to non-manager nodes] *************************\n\nTASK [Copy custom CA certificates] *********************************************\nThursday 26 March 2026  02:11:43 +0000 (0:00:01.868)       0:00:05.497 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\nTASK [Run update-ca-certificates] **********************************************\nThursday 26 March 2026  02:11:44 +0000 (0:00:01.573)       0:00:07.071 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Run update-ca-trust] *****************************************************\nThursday 26 March 2026  02:11:48 +0000 (0:00:03.701)       0:00:10.773 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Add a workaround service] ************************************************\n\nTASK [Copy workarounds.sh scripts] *********************************************\nThursday 26 March 2026  02:11:49 +0000 (0:00:00.802)       0:00:11.575 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Copy workarounds systemd unit file] **************************************\nThursday 26 March 2026  02:11:50 +0000 (0:00:01.725)       0:00:13.300 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Reload systemd daemon] ***************************************************\nThursday 26 March 2026  02:11:52 +0000 (0:00:01.832)       0:00:15.133 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable workarounds.service (Debian)] *************************************\nThursday 26 March 2026  02:11:54 +0000 (0:00:01.766)       0:00:16.900 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable and start workarounds.service (RedHat)] ***************************\nThursday 26 March 2026  02:11:56 +0000 (0:00:02.071)       0:00:18.971 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************\n\nTASK [Install python3-docker] **************************************************\nThursday 26 March 2026  02:11:57 +0000 (0:00:00.679)       0:00:19.650 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:11:59 +0000 (0:00:02.830)       0:00:22.481 ******** \n=============================================================================== \nRun update-ca-certificates ---------------------------------------------- 3.70s\nInstall python3-docker -------------------------------------------------- 2.83s\nApply netplan configuration --------------------------------------------- 2.66s\nEnable workarounds.service (Debian) ------------------------------------- 2.07s\nApply netplan configuration --------------------------------------------- 1.87s\nCopy workarounds systemd unit file -------------------------------------- 1.83s\nReload systemd daemon --------------------------------------------------- 1.77s\nCopy workarounds.sh scripts --------------------------------------------- 1.73s\nCopy custom CA certificates --------------------------------------------- 1.57s\nGroup hosts based on virtualization_role -------------------------------- 0.84s\nRun update-ca-trust ----------------------------------------------------- 0.80s\nEnable and start workarounds.service (RedHat) --------------------------- 0.68s\n+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-26 02:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 82c5765b-791d-4916-9564-ec8dce0bd1ed (reboot) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 82c5765b-791d-4916-9564-ec8dce0bd1ed (reboot) has been started and output is visible here.\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:17 +0000 (0:00:00.208)       0:00:00.208 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:17 +0000 (0:00:00.118)       0:00:00.326 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:18 +0000 (0:00:00.949)       0:00:01.276 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:18 +0000 (0:00:00.120)       0:00:01.396 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:18 +0000 (0:00:00.113)       0:00:01.510 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:19 +0000 (0:00:00.679)       0:00:02.189 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:19 +0000 (0:00:00.117)       0:00:02.306 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:19 +0000 (0:00:00.220)       0:00:02.527 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:20 +0000 (0:00:00.675)       0:00:03.203 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:20 +0000 (0:00:00.117)       0:00:03.321 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:20 +0000 (0:00:00.119)       0:00:03.440 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:21 +0000 (0:00:00.672)       0:00:04.112 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:21 +0000 (0:00:00.133)       0:00:04.246 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:21 +0000 (0:00:00.119)       0:00:04.366 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:22 +0000 (0:00:00.695)       0:00:05.062 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nThursday 26 March 2026  02:12:22 +0000 (0:00:00.125)       0:00:05.187 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nThursday 26 March 2026  02:12:22 +0000 (0:00:00.122)       0:00:05.310 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nThursday 26 March 2026  02:12:23 +0000 (0:00:00.667)       0:00:05.977 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:12:23 +0000 (0:00:00.037)       0:00:06.015 ******** \n=============================================================================== \nReboot system - do not wait for the reboot to complete ------------------ 4.34s\nExit playbook, if user did not mean to reboot systems ------------------- 0.81s\nReboot system - wait for the reboot to complete ------------------------- 0.65s\n+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-26 02:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5436eb36-cc72-4e39-ad74-719ebae5ca2c (wait-for-connection) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5436eb36-cc72-4e39-ad74-719ebae5ca2c (wait-for-connection) has been started and output is visible here.\u001b[0m\n\nPLAY [Wait until remote systems are reachable] *********************************\n\nTASK [Wait until remote system is reachable] ***********************************\nThursday 26 March 2026  02:12:40 +0000 (0:00:00.242)       0:00:00.242 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:12:52 +0000 (0:00:11.540)       0:00:11.783 ******** \n=============================================================================== \nWait until remote system is reachable ---------------------------------- 11.54s\n+ osism apply hddtemp\n\u001b[32m2026-03-26 02:13:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c3f3a815-b5d4-49d2-b3ca-aab97cf0a9af (hddtemp) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:13:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c3f3a815-b5d4-49d2-b3ca-aab97cf0a9af (hddtemp) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role hddtemp] ******************************************************\n\nTASK [osism.services.hddtemp : Gather variables for each operating system] *****\nThursday 26 March 2026  02:13:09 +0000 (0:00:00.268)       0:00:00.268 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific install tasks] ****\nThursday 26 March 2026  02:13:10 +0000 (0:00:00.792)       0:00:01.060 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Remove hddtemp package] *************************\nThursday 26 March 2026  02:13:12 +0000 (0:00:01.332)       0:00:02.393 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************\nThursday 26 March 2026  02:13:14 +0000 (0:00:02.024)       0:00:04.418 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Check if drivetemp module is available] *********\nThursday 26 March 2026  02:13:15 +0000 (0:00:01.239)       0:00:05.658 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************\nThursday 26 March 2026  02:13:16 +0000 (0:00:01.295)       0:00:06.953 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Install lm-sensors] *****************************\nThursday 26 March 2026  02:13:17 +0000 (0:00:00.935)       0:00:07.889 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific service tasks] ****\nThursday 26 March 2026  02:13:30 +0000 (0:00:12.465)       0:00:20.355 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Manage lm-sensors service] **********************\nThursday 26 March 2026  02:13:31 +0000 (0:00:01.506)       0:00:21.861 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:13:33 +0000 (0:00:01.969)       0:00:23.830 ******** \n=============================================================================== \nosism.services.hddtemp : Install lm-sensors ---------------------------- 12.47s\nosism.services.hddtemp : Remove hddtemp package ------------------------- 2.02s\nosism.services.hddtemp : Manage lm-sensors service ---------------------- 1.97s\nosism.services.hddtemp : Include distribution specific service tasks ---- 1.51s\nosism.services.hddtemp : Include distribution specific install tasks ---- 1.33s\nosism.services.hddtemp : Check if drivetemp module is available --------- 1.30s\nosism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.24s\nosism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.94s\nosism.services.hddtemp : Gather variables for each operating system ----- 0.79s\n++ semver 9.5.0 7.1.1\n+ [[ 1 -ge 0 ]]\n+ sudo systemctl restart manager.service\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\nARA in ceph-ansible already disabled.\nARA in kolla-ansible already disabled.\nARA in osism-ansible already disabled.\nARA in osism-kubernetes already disabled.\n+ osism apply gather-facts\n\u001b[32m2026-03-26 02:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2b1a1c1b-3e66-41d1-9d10-d820120a0f50 (gather-facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2b1a1c1b-3e66-41d1-9d10-d820120a0f50 (gather-facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:15:12 +0000 (0:00:00.237)       0:00:00.237 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nThursday 26 March 2026  02:15:20 +0000 (0:00:08.322)       0:00:08.559 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:15:21 +0000 (0:00:00.598)       0:00:09.158 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 8.32s\nGather facts for all hosts ---------------------------------------------- 0.60s\n+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack\n+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring\n+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack\n+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring\n+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry\n+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images\n+ [[ false == \\t\\r\\u\\e ]]",
                            "stdout_lines": [
                                "",
                                "# DEPLOY MANAGER",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY MANAGER'",
                                "+ echo",
                                "+ cat /opt/manager-vars.sh",
                                "export NUMBER_OF_NODES=6",
                                "",
                                "export CEPH_VERSION=reef",
                                "export CONFIGURATION_VERSION=main",
                                "export MANAGER_VERSION=9.5.0",
                                "export OPENSTACK_VERSION=2024.2",
                                "",
                                "export ARA=false",
                                "export DEPLOY_MODE=manager",
                                "export TEMPEST=false",
                                "export IS_ZUUL=true",
                                "",
                                "export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "export EXTERNAL_API=false",
                                "",
                                "export IMAGE_USER=ubuntu",
                                "export IMAGE_NODE_USER=ubuntu",
                                "",
                                "export CEPH_STACK=ceph-ansible",
                                "",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver",
                                "+ docker version",
                                "Client: Docker Engine - Community",
                                " Version:           27.5.1",
                                " API version:       1.47",
                                " Go version:        go1.22.11",
                                " Git commit:        9f9e405",
                                " Built:             Wed Jan 22 13:41:48 2025",
                                " OS/Arch:           linux/amd64",
                                " Context:           default",
                                "",
                                "Server: Docker Engine - Community",
                                " Engine:",
                                "  Version:          27.5.1",
                                "  API version:      1.47 (minimum version 1.24)",
                                "  Go version:       go1.22.11",
                                "  Git commit:       4c9b3b0",
                                "  Built:            Wed Jan 22 13:41:48 2025",
                                "  OS/Arch:          linux/amd64",
                                "  Experimental:     false",
                                " containerd:",
                                "  Version:          v2.2.2",
                                "  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9",
                                " runc:",
                                "  Version:          1.3.4",
                                "  GitCommit:        v1.3.4-0-gd6d73eb8",
                                " docker-init:",
                                "  Version:          0.19.0",
                                "  GitCommit:        de40ad0",
                                "+ sh -c /opt/configuration/scripts/deploy/000-manager.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-manager-version.sh 9.5.0",
                                "+ set -e",
                                "+ VERSION=9.5.0",
                                "+ sed -i 's/manager_version: .*/manager_version: 9.5.0/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i /ceph_version:/d /opt/configuration/environments/manager/configuration.yml",
                                "+ sed -i /openstack_version:/d /opt/configuration/environments/manager/configuration.yml",
                                "+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh",
                                "/opt/configuration ~",
                                "+ set -e",
                                "+ pushd /opt/configuration",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging",
                                "Requirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)",
                                "Requirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)",
                                "Requirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)",
                                "Requirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)",
                                "Requirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)",
                                "Requirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)",
                                "Requirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)",
                                "Requirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)",
                                "Requirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)",
                                "Requirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)",
                                "Requirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)",
                                "Requirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)",
                                "Requirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)",
                                "Requirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)",
                                "++ which gilt",
                                "+ GILT=/opt/venv/bin/gilt",
                                "+ /opt/venv/bin/gilt overlay",
                                "osism.cfg-generics:",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/",
                                "  - running `rm render-images.py` in /opt/configuration/environments/manager/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/",
                                "  - running `rm set-versions.py` in /opt/configuration/environments/",
                                "~",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ popd",
                                "+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ ceph-ansible == \\r\\o\\o\\k ]]",
                                "++ semver 9.5.0 7.0.0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ echo 'enable_osism_kubernetes: true'",
                                "++ semver 9.5.0 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 2024.2 2025.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ /opt/configuration/scripts/enable-resource-nodes.sh",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml",
                                "",
                                "PLAY [Copy custom facts] *******************************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY [Before the deployment of the manager] ************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get /opt/manager-vars.sh] ************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add ara_server_mariadb_volume_type parameter] ****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Add netbox_enable parameter] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Check if /etc/OTC_region exist] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add nova_compute_virt_type parameter] ************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Apply role traefik] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply traefik role] ******************************************************",
                                "included: osism.services.traefik for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/traefik)",
                                "changed: [testbed-manager] => (item=/opt/traefik/certificates)",
                                "changed: [testbed-manager] => (item=/opt/traefik/configuration)",
                                "",
                                "TASK [osism.services.traefik : Copy configuration files] ***********************",
                                "changed: [testbed-manager] => (item=traefik.yml)",
                                "changed: [testbed-manager] => (item=traefik.env)",
                                "changed: [testbed-manager] => (item=certificates.yml)",
                                "",
                                "TASK [osism.services.traefik : Copy certificate cert files] ********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy certificate key files] *********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy dynamic configuration] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Remove dynamic configuration] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create traefik external network] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Manage traefik service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.traefik : Restart traefik service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Deploy manager service] **************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply manager role] ******************************************************",
                                "included: osism.services.manager for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/ansible)",
                                "changed: [testbed-manager] => (item=/opt/archive)",
                                "changed: [testbed-manager] => (item=/opt/manager/configuration)",
                                "changed: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "changed: [testbed-manager] => (item=/opt/manager/secrets)",
                                "changed: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "changed: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "changed: [testbed-manager] => (item=ara)",
                                "changed: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "changed: [testbed-manager] => (item=conductor)",
                                "changed: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "changed: [testbed-manager] => (item=osism)",
                                "changed: [testbed-manager] => (item=osism-update-docker)",
                                "changed: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "changed: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:0.20251130.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:0.20251130.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:0.20251130.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:0.20251130.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:0.20251130.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:0.20251130.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:0.20251130.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:0.20251130.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:0.20251130.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:0.20251130.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251130.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   ",
                                "",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                         COMMAND                  SERVICE                CREATED         STATUS                        PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:0.20251130.0           \"/entrypoint.sh osis\u2026\"   ceph-ansible           2 minutes ago   Up About a minute (healthy)   ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:0.20251130.0          \"/entrypoint.sh osis\u2026\"   kolla-ansible          2 minutes ago   Up About a minute (healthy)   ",
                                "manager-api-1                    registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago   Up 2 minutes (healthy)        192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                    \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago   Up 2 minutes (healthy)        8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:0.20251130.0   \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   2 minutes ago   Up About a minute (healthy)   ",
                                "manager-listener-1               registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4          \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago   Up 2 minutes (healthy)        3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine      \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago   Up 2 minutes (healthy)        6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:0.20251130.0          \"/entrypoint.sh osis\u2026\"   osism-ansible          2 minutes ago   Up About a minute (healthy)   ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:0.20251130.1         \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago   Up 2 minutes                  192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:0.20251130.0       \"/entrypoint.sh osis\u2026\"   osism-kubernetes       2 minutes ago   Up About a minute (healthy)   ",
                                "osismclient                      registry.osism.tech/osism/osism:0.20251130.1                  \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago   Up 2 minutes (healthy)        ",
                                "++ semver 9.5.0 7.0.0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg",
                                "+ osism apply resolvconf -l testbed-manager",
                                "\u001b[32m2026-03-26 01:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b338eb90-9c5b-4feb-973b-e917f7311fb8 (resolvconf) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b338eb90-9c5b-4feb-973b-e917f7311fb8 (resolvconf) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role resolvconf] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Thursday 26 March 2026  01:55:17 +0000 (0:00:00.157)       0:00:00.157 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Thursday 26 March 2026  01:55:21 +0000 (0:00:03.974)       0:00:04.131 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Thursday 26 March 2026  01:55:21 +0000 (0:00:00.071)       0:00:04.203 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Thursday 26 March 2026  01:55:21 +0000 (0:00:00.091)       0:00:04.294 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Thursday 26 March 2026  01:55:21 +0000 (0:00:00.081)       0:00:04.376 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Thursday 26 March 2026  01:55:22 +0000 (0:00:01.247)       0:00:05.623 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Thursday 26 March 2026  01:55:22 +0000 (0:00:00.066)       0:00:05.690 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Thursday 26 March 2026  01:55:23 +0000 (0:00:00.528)       0:00:06.218 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Thursday 26 March 2026  01:55:23 +0000 (0:00:00.091)       0:00:06.310 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Thursday 26 March 2026  01:55:23 +0000 (0:00:00.562)       0:00:06.872 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Thursday 26 March 2026  01:55:25 +0000 (0:00:01.131)       0:00:08.004 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Thursday 26 March 2026  01:55:26 +0000 (0:00:01.017)       0:00:09.021 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Thursday 26 March 2026  01:55:26 +0000 (0:00:00.084)       0:00:09.106 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  01:55:27 +0000 (0:00:01.180)       0:00:10.286 ******** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.97s",
                                "osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.25s",
                                "osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.18s",
                                "osism.commons.resolvconf : Copy configuration files --------------------- 1.13s",
                                "osism.commons.resolvconf : Start/enable systemd-resolved service -------- 1.02s",
                                "osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.56s",
                                "osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.53s",
                                "osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.09s",
                                "osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s",
                                "osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s",
                                "osism.commons.resolvconf : Include distribution specific installation tasks --- 0.08s",
                                "osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s",
                                "osism.commons.resolvconf : Install package systemd-resolved ------------- 0.07s",
                                "+ osism apply sshconfig",
                                "\u001b[32m2026-03-26 01:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c0f329a2-f521-4b7e-839d-4cec16d416a2 (sshconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c0f329a2-f521-4b7e-839d-4cec16d416a2 (sshconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role sshconfig] ****************************************************",
                                "",
                                "TASK [osism.commons.sshconfig : Get home directory of operator user] ***********",
                                "Thursday 26 March 2026  01:55:44 +0000 (0:00:00.169)       0:00:00.169 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************",
                                "Thursday 26 March 2026  01:55:44 +0000 (0:00:00.551)       0:00:00.721 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure config for each host exist] *************",
                                "Thursday 26 March 2026  01:55:45 +0000 (0:00:00.563)       0:00:01.284 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Add extra config] ******************************",
                                "Thursday 26 March 2026  01:55:51 +0000 (0:00:06.025)       0:00:07.310 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Assemble ssh config] ***************************",
                                "Thursday 26 March 2026  01:55:51 +0000 (0:00:00.088)       0:00:07.398 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  01:55:52 +0000 (0:00:00.631)       0:00:08.029 ******** ",
                                "=============================================================================== ",
                                "osism.commons.sshconfig : Ensure config for each host exist ------------- 6.03s",
                                "osism.commons.sshconfig : Assemble ssh config --------------------------- 0.63s",
                                "osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.56s",
                                "osism.commons.sshconfig : Get home directory of operator user ----------- 0.55s",
                                "osism.commons.sshconfig : Add extra config ------------------------------ 0.09s",
                                "+ osism apply known-hosts",
                                "\u001b[32m2026-03-26 01:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1fa85f69-74d9-41fc-88eb-077113f41686 (known-hosts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1fa85f69-74d9-41fc-88eb-077113f41686 (known-hosts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role known_hosts] **************************************************",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***",
                                "Thursday 26 March 2026  01:56:09 +0000 (0:00:00.179)       0:00:00.179 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***",
                                "Thursday 26 March 2026  01:56:15 +0000 (0:00:06.471)       0:00:06.651 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:15 +0000 (0:00:00.175)       0:00:06.827 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJ02fclj3yBlaRFCMRlBqzxa6JpoNn0+0cesF8DzED/Y)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCjl441PvBESBlyBRx1uu1H4fE7tRu8wdi33Q8JFqeLTaQZjzDEdIz6uCS4P7vly4E/qLLf8bB/8UcNRyQNcICQGP4n1FNR/EKvBi3/Gyc8AFEHmafJBMQ+fPJKZT16DdkyJaug5SebtzLWoBu55VpcJBQEuazelbRRNG8Xz6iBrkfu/Kec/g0Rno6KOcdUjXqp9eXijo2MLaMUS8xKb9xvjzA8lG94Nfv5kLJxesAhlOo3+QdD7zaeMZ7NUtCfAV0/o8Ruvk/OCfm+9xqBh34HvSOpZATYXPchqL5K63v6jn0PFjt3YFZ0gIUmA1WSFQpb6laHtVAv7J1OW2xdYo6gotU1Cr37/4ig+HHn9P+oQgsmR48E/LQGDOE8e2lb9vbBc00XUT6BPP5pLf9JyRJiE8oeTuYbPu7EAcxRe0Ub7IiJPGMHrzRYtxLMNbIcIC3Albt9+yYod/l1CB/xtv2cBu/ChuADP3Zn4908Xfv7x+enwcnvCi1DiX8TxkEolms=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBP5uNsSXygIZlJfKHbiEnkj18urKp5QMG+27OXFSyVeZCLFzqulV6Rzhv/iPWmiLWxR1gs0vX6AnWLfHyOAmsI4=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:17 +0000 (0:00:01.243)       0:00:08.071 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC7MP4ethAPLioYMXHsztWTZ5ixht+3Sq7KMlb6HBjt4WWGu5avxBzCqnNkYFhF7TM+NZp4FPbph0s1vt1MbCjwAX101M9p/xXyIg6HqiULDQGPnSM9m+LRiEsrASOQGF6aFihMRdcJRhnREJZq9qGdmItuWj9KhGkPm8uRtSFbq+buvlAumNVyRGX/2g9hRT6OPPaBs/0vnzqWm0PY24uelp14behZeTqbe67Xef6Ejb2CPSpASVYAK7cBu4QLBtZt+ComHffq2pp5kC/BZzYHIkNSsmHez8A+uBRCSgkLTqIoCB1KlEDd4YUFkUqRwLSO3P+YeXwrdsuLV64aUMYp+mx/hn1UhZf52iPN5/4dhu/u1pNdyfBgkAI8YresJOdSH6Ztn91oOS0NfNvirYGQGgZQ8F6GAHw/Jqj4Tsu0/0pBD6p34E7ve6hpBikPhSLPtsJmfT5TuU1FdhBMtq6IGmYZfCa24W3TuSRsZnpNeoagjyEJOIujgECw1Gl5Vfc=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEJBAfKNBqc6hZaPCToSyiH071v/szTHinZTYmvWtQiYY2BD9OmGZbDU8lB1KJXImA4sH06NV8BnsqYFdWToJUQ=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHe8Qmub5XmBttmFEUBkU1KnPXMx15s5oXemfYKWwqe/)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:18 +0000 (0:00:01.203)       0:00:09.274 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+4xLf8E9LJe8aia/EpM0v0rX+8sflF8vLZvkHgEXWEhHuSqtxzIL8ACbdmnHD33O2W0UBsUAyKb19kCZTvLXTqmZ2+J24oE3ZFXc8YL5hTd5Yl4TRhBUjg9m31EizACs/0jAMW2ZKSmiAC0t9dhTj2ZMMJ9YMi/kGEhZ1c3/HNB0H3V8krwQERPfDvu8cEHaZVKLA2E86hUr7O4LpXVaMjYarBwX2WTY5YK2l3fkb7rJAigD7P858LeetHWGn6w4gw6illPyvLV50RMhF4BD/KdhvdPV2Zilyh/H23zhuvOy1he82obLlwFNK7mibsaxgEJrsx0VpcFsA6bygql94UvQR3dKTb4Rc34++oxIQgdFsdrIKHadlhGcaEekjU6ZI0BKd0qE511yymd1N3ooYoVHbf8mhhr1oKWq/os46wBLhEyuyNYD3xS2L/YX2jFpKgiU4PIYyJZy+WDIkiuZNjucyKTdIci0+wIJYghYdt9ZWtDIyBEJn6QjDxRRYWaE=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLOteKbqWmsecSgYs+xqbb6nNATEUQK4q13kbcpg1b1i+s0qETmOp+d8ADubOahpy9BafeOlLLvip9rzRVjYfSs=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICb08xQfqtFt6WLy7uNMC8Hgb7gIH7sGbdc6Gp7j5cy2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:19 +0000 (0:00:01.208)       0:00:10.482 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDM4uc4YzFGwjqMlk336TEFW9Sr5NydBFeDixkdguq0xoKZ7Av9HzodJRVzGN627myZhV4E815uWI3x9reyYZn0RcrcDzleefzPfPoWxct0zKx0MfT2zwMJ22JNwOTn6REmWHbevXA5ITeutZAPxJdPrwoS8s1a1Buj3Ky0tfIaCqCRAkIV5P1pJRFSlt7gZL0ImYnlZoIDfKZ/cBT9BzsRT8CWHHY37KfeVwJm37huH9rMBh/eHolG5yLNXGvKrTq0QKC1mMY2qK0APt4GmFl8KPTnNLajycV5GheWblGibpQGFM93+2Z/WqnRJf82EIxu8yCqPo3/1FhS5lFWtQRK9+I/5G7KpU3S/WCcAmUn+w2yg1uU3CAEV+RycYxvp2lqJTbwlZx/pTBotJ9R7cUBQ5aW0Nsvp8aytFY0HRc4Xa8z0Pg6UlJTSdG0IENfARZSKYkJ1g+2JRpe/lOoXOqpWEA4KW+XdPodGxf0D2YqCd/rGxlBauwYYWbZ3Imo7ic=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAubqEkG0vhsyENkRhNUdy8+oCOQJ2VjqsD1HLNi3GC9S06/AsvC5szSrOM8xPq6YUCb0ZAm8wLM6tpF+FpQCkI=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIILAw9GUy2vpfwElSJ1fGmtyFuqmYW2r9XbGM4GRo7eY)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:20 +0000 (0:00:01.216)       0:00:11.699 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ5PAiQENk+ItZkO7lCzckgIds0Q2WgCbLVLzPfl9J/MU02VMGGoMDDgddubcewkixXkbNvIDK1stUBpmzi0L0j6dHcUMdFvRgqm8GCoMsXNrfHW/ct+73eNQxULwT9bQllXOZr2nB7MltedCplvwjiFJBdmFrGAIp+GrXL0PMApIE/dd1f1qPoBloRBPpdJX1CjF1OGaQsZ4qpz7WpJrlw7thtAHdcylkEbcstauBnNgTubYjVaIo2uIJXiCWNIL0fPhNIHAMKddURALWDzhs+Xh/MW65iFiI8YfPERzwYogm9ZGXr3rX08YWezcrnK5zImuNAFUMNYrmHbLAjzsIRCewlKYooBB2dtd3+XDSqNrvX7+tVW6K8NlU917X/WU52YVtcXFisGbVrw/elyi1iLd7EwmJ9QCUMobcLT0ImLsPJAR5WpjsxApgqPjXLiRWoBFIT4ZJh3VGUe5BPDf8dT/Ij1IENkwR4wNsSm+8yKbokryeCpNJ/83tRYElNd8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKbUTEgC6bw2kPbQ69BQKKbddt6FzzQXjl7gPSrh85KeKra/4AOp2yhnEkwPsOgQiMMH5NRrWQXsymB0y+9b1Q8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGuDGp4QP0xBuwhTX5o6OTk2Lr58UqgVCuxQtBHOizHi)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:21 +0000 (0:00:01.175)       0:00:12.875 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNUp+hESk7uzg35hbkvvf+NUU3AkgbCjVB2Zkr8YJaR+BFh9GPPoZZsfxubkFFdXWW3nEN+8mFY1PIXNSyE0jYU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCtvM026ThskiZdDILB9UjPU47PsDHtiHncBI7eSm7qoZ/vFmGYxPZa4NYpRQd13zEH+bhVZEQoiEJjSXZTdefjqImCiIVEtQYLak7OohLyh+btJqEp4MZX1p+Ef87NJt6iZXTmRO9hJa7iuKN5UU6QTJuvgvy/KEw7KR/yqIwc90ds6MtpKtAKEnqodONOccyVKbdgfAngkmZ2uye4IM8jrwODbjTbvmztDxJ3Sp24S7z173HSnLssG6zn5hxZxlMJQPsshb1LZB7JEF6lISzJIVNgyjszZc+PJPIXOQ0eOfc5gVES9pBSljtremJUbyNhJTxEYrER41dO5QupDhOo2kWeNPdHCnP2TzghsdM8MYFFRQrxKuPY9/o9ba8oW0xMHcVSQiUXpxZ0envlNeCzmYATs6z4y/XPHjA1K+boWRjzlBpTOsPyxFYR8NWg16MV5V4YwxIGBkMwr86umV4mJ+4qP03CktNELRx8hMDTGQ1vt5H3sjQnWKAb5bwbJnk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC/uIeNDlNnWqqbf4nQ1iFXm8xqmTno9JV+7U0mLmoBx)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:22 +0000 (0:00:01.136)       0:00:14.011 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCgKIZ+Izl5D9pHv6Kxhr5AyZEEKn9/BK9MNhQb/QhaO78ATvWpiJlyn30Si4IfRm0v7dCgB0sW8NB5CvAGq/ZGVQE3hr4tNBZrraVdJW4ItePEAsUdTBKwaGp9SU7bwlkZDWNr4EKmSRIefON4mC+Ybn4LJccP5VfAIqMv7w/fcREx9xmgnP7OprQLP3T35T1ObQBXrDu4nOSoxOUa3eeTBzb0M2JXK31R8MAkfBpl3xt/3e/97jB5oOcK9t9kbC3QwZmIYAcx+XCSmxyQ7kcR5PTtZoBVwzPXqPMU0tCc0UXDEJRnsTrXIJ0wArclmDuuIyLundoxRn57nEmHByjdHzJ5u/RqZYNJHs0LWyPDC7FOQ2ek+xnsnzjizES2z96sVFk385BUNTp+/Iv74/8jOP7Glx2+Q3TQc5BCM45SISAWv/C65y24IJLwG5hfzk0+g1I/moUPPDThzsC9weUzFs4C0dDIYlckw90A0PCUIFSGzaLQtti2t3GTfb6oISU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK2OXpbGe9PH2FJGGvIa5N+853ic5OP4B+gsXUDRcJHyzkcI+mMK2k9rDDpbgobH0QeLp+WUIjUIHSmvkdzCtC4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIK2b9SZZ/SJ6wJY/vZIQL6KvWIfJRQxK2nxhYfFkytaC)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***",
                                "Thursday 26 March 2026  01:56:24 +0000 (0:00:01.211)       0:00:15.223 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***",
                                "Thursday 26 March 2026  01:56:29 +0000 (0:00:05.758)       0:00:20.981 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:30 +0000 (0:00:00.184)       0:00:21.166 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCjl441PvBESBlyBRx1uu1H4fE7tRu8wdi33Q8JFqeLTaQZjzDEdIz6uCS4P7vly4E/qLLf8bB/8UcNRyQNcICQGP4n1FNR/EKvBi3/Gyc8AFEHmafJBMQ+fPJKZT16DdkyJaug5SebtzLWoBu55VpcJBQEuazelbRRNG8Xz6iBrkfu/Kec/g0Rno6KOcdUjXqp9eXijo2MLaMUS8xKb9xvjzA8lG94Nfv5kLJxesAhlOo3+QdD7zaeMZ7NUtCfAV0/o8Ruvk/OCfm+9xqBh34HvSOpZATYXPchqL5K63v6jn0PFjt3YFZ0gIUmA1WSFQpb6laHtVAv7J1OW2xdYo6gotU1Cr37/4ig+HHn9P+oQgsmR48E/LQGDOE8e2lb9vbBc00XUT6BPP5pLf9JyRJiE8oeTuYbPu7EAcxRe0Ub7IiJPGMHrzRYtxLMNbIcIC3Albt9+yYod/l1CB/xtv2cBu/ChuADP3Zn4908Xfv7x+enwcnvCi1DiX8TxkEolms=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBP5uNsSXygIZlJfKHbiEnkj18urKp5QMG+27OXFSyVeZCLFzqulV6Rzhv/iPWmiLWxR1gs0vX6AnWLfHyOAmsI4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJ02fclj3yBlaRFCMRlBqzxa6JpoNn0+0cesF8DzED/Y)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:31 +0000 (0:00:01.113)       0:00:22.279 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC7MP4ethAPLioYMXHsztWTZ5ixht+3Sq7KMlb6HBjt4WWGu5avxBzCqnNkYFhF7TM+NZp4FPbph0s1vt1MbCjwAX101M9p/xXyIg6HqiULDQGPnSM9m+LRiEsrASOQGF6aFihMRdcJRhnREJZq9qGdmItuWj9KhGkPm8uRtSFbq+buvlAumNVyRGX/2g9hRT6OPPaBs/0vnzqWm0PY24uelp14behZeTqbe67Xef6Ejb2CPSpASVYAK7cBu4QLBtZt+ComHffq2pp5kC/BZzYHIkNSsmHez8A+uBRCSgkLTqIoCB1KlEDd4YUFkUqRwLSO3P+YeXwrdsuLV64aUMYp+mx/hn1UhZf52iPN5/4dhu/u1pNdyfBgkAI8YresJOdSH6Ztn91oOS0NfNvirYGQGgZQ8F6GAHw/Jqj4Tsu0/0pBD6p34E7ve6hpBikPhSLPtsJmfT5TuU1FdhBMtq6IGmYZfCa24W3TuSRsZnpNeoagjyEJOIujgECw1Gl5Vfc=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEJBAfKNBqc6hZaPCToSyiH071v/szTHinZTYmvWtQiYY2BD9OmGZbDU8lB1KJXImA4sH06NV8BnsqYFdWToJUQ=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHe8Qmub5XmBttmFEUBkU1KnPXMx15s5oXemfYKWwqe/)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:32 +0000 (0:00:01.189)       0:00:23.469 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+4xLf8E9LJe8aia/EpM0v0rX+8sflF8vLZvkHgEXWEhHuSqtxzIL8ACbdmnHD33O2W0UBsUAyKb19kCZTvLXTqmZ2+J24oE3ZFXc8YL5hTd5Yl4TRhBUjg9m31EizACs/0jAMW2ZKSmiAC0t9dhTj2ZMMJ9YMi/kGEhZ1c3/HNB0H3V8krwQERPfDvu8cEHaZVKLA2E86hUr7O4LpXVaMjYarBwX2WTY5YK2l3fkb7rJAigD7P858LeetHWGn6w4gw6illPyvLV50RMhF4BD/KdhvdPV2Zilyh/H23zhuvOy1he82obLlwFNK7mibsaxgEJrsx0VpcFsA6bygql94UvQR3dKTb4Rc34++oxIQgdFsdrIKHadlhGcaEekjU6ZI0BKd0qE511yymd1N3ooYoVHbf8mhhr1oKWq/os46wBLhEyuyNYD3xS2L/YX2jFpKgiU4PIYyJZy+WDIkiuZNjucyKTdIci0+wIJYghYdt9ZWtDIyBEJn6QjDxRRYWaE=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLOteKbqWmsecSgYs+xqbb6nNATEUQK4q13kbcpg1b1i+s0qETmOp+d8ADubOahpy9BafeOlLLvip9rzRVjYfSs=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICb08xQfqtFt6WLy7uNMC8Hgb7gIH7sGbdc6Gp7j5cy2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:33 +0000 (0:00:01.183)       0:00:24.653 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAubqEkG0vhsyENkRhNUdy8+oCOQJ2VjqsD1HLNi3GC9S06/AsvC5szSrOM8xPq6YUCb0ZAm8wLM6tpF+FpQCkI=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIILAw9GUy2vpfwElSJ1fGmtyFuqmYW2r9XbGM4GRo7eY)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDM4uc4YzFGwjqMlk336TEFW9Sr5NydBFeDixkdguq0xoKZ7Av9HzodJRVzGN627myZhV4E815uWI3x9reyYZn0RcrcDzleefzPfPoWxct0zKx0MfT2zwMJ22JNwOTn6REmWHbevXA5ITeutZAPxJdPrwoS8s1a1Buj3Ky0tfIaCqCRAkIV5P1pJRFSlt7gZL0ImYnlZoIDfKZ/cBT9BzsRT8CWHHY37KfeVwJm37huH9rMBh/eHolG5yLNXGvKrTq0QKC1mMY2qK0APt4GmFl8KPTnNLajycV5GheWblGibpQGFM93+2Z/WqnRJf82EIxu8yCqPo3/1FhS5lFWtQRK9+I/5G7KpU3S/WCcAmUn+w2yg1uU3CAEV+RycYxvp2lqJTbwlZx/pTBotJ9R7cUBQ5aW0Nsvp8aytFY0HRc4Xa8z0Pg6UlJTSdG0IENfARZSKYkJ1g+2JRpe/lOoXOqpWEA4KW+XdPodGxf0D2YqCd/rGxlBauwYYWbZ3Imo7ic=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:34 +0000 (0:00:01.177)       0:00:25.830 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKbUTEgC6bw2kPbQ69BQKKbddt6FzzQXjl7gPSrh85KeKra/4AOp2yhnEkwPsOgQiMMH5NRrWQXsymB0y+9b1Q8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGuDGp4QP0xBuwhTX5o6OTk2Lr58UqgVCuxQtBHOizHi)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ5PAiQENk+ItZkO7lCzckgIds0Q2WgCbLVLzPfl9J/MU02VMGGoMDDgddubcewkixXkbNvIDK1stUBpmzi0L0j6dHcUMdFvRgqm8GCoMsXNrfHW/ct+73eNQxULwT9bQllXOZr2nB7MltedCplvwjiFJBdmFrGAIp+GrXL0PMApIE/dd1f1qPoBloRBPpdJX1CjF1OGaQsZ4qpz7WpJrlw7thtAHdcylkEbcstauBnNgTubYjVaIo2uIJXiCWNIL0fPhNIHAMKddURALWDzhs+Xh/MW65iFiI8YfPERzwYogm9ZGXr3rX08YWezcrnK5zImuNAFUMNYrmHbLAjzsIRCewlKYooBB2dtd3+XDSqNrvX7+tVW6K8NlU917X/WU52YVtcXFisGbVrw/elyi1iLd7EwmJ9QCUMobcLT0ImLsPJAR5WpjsxApgqPjXLiRWoBFIT4ZJh3VGUe5BPDf8dT/Ij1IENkwR4wNsSm+8yKbokryeCpNJ/83tRYElNd8=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:35 +0000 (0:00:01.199)       0:00:27.030 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCtvM026ThskiZdDILB9UjPU47PsDHtiHncBI7eSm7qoZ/vFmGYxPZa4NYpRQd13zEH+bhVZEQoiEJjSXZTdefjqImCiIVEtQYLak7OohLyh+btJqEp4MZX1p+Ef87NJt6iZXTmRO9hJa7iuKN5UU6QTJuvgvy/KEw7KR/yqIwc90ds6MtpKtAKEnqodONOccyVKbdgfAngkmZ2uye4IM8jrwODbjTbvmztDxJ3Sp24S7z173HSnLssG6zn5hxZxlMJQPsshb1LZB7JEF6lISzJIVNgyjszZc+PJPIXOQ0eOfc5gVES9pBSljtremJUbyNhJTxEYrER41dO5QupDhOo2kWeNPdHCnP2TzghsdM8MYFFRQrxKuPY9/o9ba8oW0xMHcVSQiUXpxZ0envlNeCzmYATs6z4y/XPHjA1K+boWRjzlBpTOsPyxFYR8NWg16MV5V4YwxIGBkMwr86umV4mJ+4qP03CktNELRx8hMDTGQ1vt5H3sjQnWKAb5bwbJnk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNUp+hESk7uzg35hbkvvf+NUU3AkgbCjVB2Zkr8YJaR+BFh9GPPoZZsfxubkFFdXWW3nEN+8mFY1PIXNSyE0jYU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC/uIeNDlNnWqqbf4nQ1iFXm8xqmTno9JV+7U0mLmoBx)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Thursday 26 March 2026  01:56:37 +0000 (0:00:01.197)       0:00:28.228 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCgKIZ+Izl5D9pHv6Kxhr5AyZEEKn9/BK9MNhQb/QhaO78ATvWpiJlyn30Si4IfRm0v7dCgB0sW8NB5CvAGq/ZGVQE3hr4tNBZrraVdJW4ItePEAsUdTBKwaGp9SU7bwlkZDWNr4EKmSRIefON4mC+Ybn4LJccP5VfAIqMv7w/fcREx9xmgnP7OprQLP3T35T1ObQBXrDu4nOSoxOUa3eeTBzb0M2JXK31R8MAkfBpl3xt/3e/97jB5oOcK9t9kbC3QwZmIYAcx+XCSmxyQ7kcR5PTtZoBVwzPXqPMU0tCc0UXDEJRnsTrXIJ0wArclmDuuIyLundoxRn57nEmHByjdHzJ5u/RqZYNJHs0LWyPDC7FOQ2ek+xnsnzjizES2z96sVFk385BUNTp+/Iv74/8jOP7Glx2+Q3TQc5BCM45SISAWv/C65y24IJLwG5hfzk0+g1I/moUPPDThzsC9weUzFs4C0dDIYlckw90A0PCUIFSGzaLQtti2t3GTfb6oISU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK2OXpbGe9PH2FJGGvIa5N+853ic5OP4B+gsXUDRcJHyzkcI+mMK2k9rDDpbgobH0QeLp+WUIjUIHSmvkdzCtC4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIK2b9SZZ/SJ6wJY/vZIQL6KvWIfJRQxK2nxhYfFkytaC)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write static known_hosts entries] ************",
                                "Thursday 26 March 2026  01:56:38 +0000 (0:00:01.261)       0:00:29.489 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write extra known_hosts entries] *************",
                                "Thursday 26 March 2026  01:56:38 +0000 (0:00:00.166)       0:00:29.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Delete known_hosts entries] ******************",
                                "Thursday 26 March 2026  01:56:38 +0000 (0:00:00.071)       0:00:29.728 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Set file permissions] ************************",
                                "Thursday 26 March 2026  01:56:38 +0000 (0:00:00.066)       0:00:29.794 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  01:56:39 +0000 (0:00:00.807)       0:00:30.602 ******** ",
                                "=============================================================================== ",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 6.47s",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.76s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.26s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.24s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.22s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.21s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.21s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.20s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.19s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.18s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.14s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.11s",
                                "osism.commons.known_hosts : Set file permissions ------------------------ 0.81s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.18s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.18s",
                                "osism.commons.known_hosts : Write static known_hosts entries ------------ 0.17s",
                                "+ osism apply squid",
                                "\u001b[32m2026-03-26 01:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 22181c16-c6bd-4755-9a52-705156a3718e (squid) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 22181c16-c6bd-4755-9a52-705156a3718e (squid) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role squid] ********************************************************",
                                "",
                                "TASK [osism.services.squid : Include install tasks] ****************************",
                                "Thursday 26 March 2026  01:56:56 +0000 (0:00:00.184)       0:00:00.184 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Install required packages] ************************",
                                "Thursday 26 March 2026  01:56:56 +0000 (0:00:00.108)       0:00:00.293 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Create required directories] **********************",
                                "Thursday 26 March 2026  01:56:58 +0000 (0:00:01.915)       0:00:02.208 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy squid configuration files] *******************",
                                "Thursday 26 March 2026  01:57:00 +0000 (0:00:01.319)       0:00:03.528 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***",
                                "Thursday 26 March 2026  01:57:01 +0000 (0:00:01.148)       0:00:04.676 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy docker-compose.yml file] *********************",
                                "Thursday 26 March 2026  01:57:01 +0000 (0:00:00.386)       0:00:05.062 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Manage squid service] *****************************",
                                "Thursday 26 March 2026  01:57:02 +0000 (0:00:00.961)       0:00:06.023 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Restart squid service] *****************",
                                "Thursday 26 March 2026  01:57:39 +0000 (0:00:36.629)       0:00:42.653 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******",
                                "Thursday 26 March 2026  01:57:51 +0000 (0:00:12.112)       0:00:54.765 ******** ",
                                "Pausing for 60 seconds",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***",
                                "Thursday 26 March 2026  01:58:51 +0000 (0:01:00.088)       0:01:54.854 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****",
                                "Thursday 26 March 2026  01:58:51 +0000 (0:00:00.074)       0:01:54.928 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  01:58:52 +0000 (0:00:00.567)       0:01:55.495 ******** ",
                                "=============================================================================== ",
                                "osism.services.squid : Wait for squid service to start ----------------- 60.09s",
                                "osism.services.squid : Manage squid service ---------------------------- 36.63s",
                                "osism.services.squid : Restart squid service --------------------------- 12.11s",
                                "osism.services.squid : Install required packages ------------------------ 1.92s",
                                "osism.services.squid : Create required directories ---------------------- 1.32s",
                                "osism.services.squid : Copy squid configuration files ------------------- 1.15s",
                                "osism.services.squid : Copy docker-compose.yml file --------------------- 0.96s",
                                "osism.services.squid : Wait for an healthy squid service ---------------- 0.57s",
                                "osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.39s",
                                "osism.services.squid : Include install tasks ---------------------------- 0.11s",
                                "osism.services.squid : Register that squid service was restarted -------- 0.07s",
                                "+ [[ 9.5.0 != \\l\\a\\t\\e\\s\\t ]]",
                                "++ semver 9.5.0 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release",
                                "+ set -e",
                                "+ NAMESPACE=kolla/release",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "++ semver 9.5.0 9.0.0",
                                "+ [[ 1 -lt 0 ]]",
                                "+ osism apply operator -u ubuntu -l testbed-nodes",
                                "\u001b[32m2026-03-26 01:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b14432b8-054d-4f4f-bffe-fa6ab43829d4 (operator) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b14432b8-054d-4f4f-bffe-fa6ab43829d4 (operator) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Make ssh pipelining working] *********************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Thursday 26 March 2026  01:59:08 +0000 (0:00:00.152)       0:00:00.152 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Do not require tty for all users] ****************************************",
                                "Thursday 26 March 2026  01:59:12 +0000 (0:00:03.324)       0:00:03.477 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply role operator] *****************************************************",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "Thursday 26 March 2026  01:59:12 +0000 (0:00:00.753)       0:00:04.230 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "Thursday 26 March 2026  01:59:13 +0000 (0:00:00.195)       0:00:04.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "Thursday 26 March 2026  01:59:13 +0000 (0:00:00.195)       0:00:04.621 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "Thursday 26 March 2026  01:59:13 +0000 (0:00:00.624)       0:00:05.246 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "Thursday 26 March 2026  01:59:14 +0000 (0:00:00.874)       0:00:06.120 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "Thursday 26 March 2026  01:59:15 +0000 (0:00:01.247)       0:00:07.367 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "Thursday 26 March 2026  01:59:17 +0000 (0:00:01.281)       0:00:08.648 ******** ",
                                "\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m",
                                "\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m",
                                "\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "Thursday 26 March 2026  01:59:18 +0000 (0:00:01.301)       0:00:09.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "Thursday 26 March 2026  01:59:18 +0000 (0:00:00.198)       0:00:10.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "Thursday 26 March 2026  01:59:18 +0000 (0:00:00.187)       0:00:10.336 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "Thursday 26 March 2026  01:59:19 +0000 (0:00:00.585)       0:00:10.921 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "Thursday 26 March 2026  01:59:19 +0000 (0:00:00.184)       0:00:11.105 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "Thursday 26 March 2026  01:59:20 +0000 (0:00:00.722)       0:00:11.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "Thursday 26 March 2026  01:59:20 +0000 (0:00:00.177)       0:00:12.006 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "Thursday 26 March 2026  01:59:20 +0000 (0:00:00.177)       0:00:12.183 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "Thursday 26 March 2026  01:59:20 +0000 (0:00:00.168)       0:00:12.352 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "Thursday 26 March 2026  01:59:21 +0000 (0:00:00.650)       0:00:13.003 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  01:59:21 +0000 (0:00:00.269)       0:00:13.273 ******** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.32s",
                                "osism.commons.operator : Set language variables in .bashrc configuration file --- 1.30s",
                                "osism.commons.operator : Copy user sudoers file ------------------------- 1.28s",
                                "osism.commons.operator : Add user to additional groups ------------------ 1.25s",
                                "osism.commons.operator : Create user ------------------------------------ 0.87s",
                                "Do not require tty for all users ---------------------------------------- 0.75s",
                                "osism.commons.operator : Set ssh authorized keys ------------------------ 0.72s",
                                "osism.commons.operator : Set password ----------------------------------- 0.65s",
                                "osism.commons.operator : Create operator group -------------------------- 0.62s",
                                "osism.commons.operator : Create .ssh directory -------------------------- 0.59s",
                                "osism.commons.operator : Unset & lock password -------------------------- 0.27s",
                                "osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.20s",
                                "osism.commons.operator : Gather variables for each operating system ----- 0.20s",
                                "osism.commons.operator : Set operator_groups variable to default value --- 0.20s",
                                "osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.19s",
                                "osism.commons.operator : Check number of SSH authorized keys ------------ 0.18s",
                                "osism.commons.operator : Delete ssh authorized keys --------------------- 0.18s",
                                "osism.commons.operator : Set authorized GitHub accounts ----------------- 0.18s",
                                "osism.commons.operator : Delete authorized GitHub accounts -------------- 0.17s",
                                "+ osism apply --environment custom facts",
                                "\u001b[32m2026-03-26 01:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m",
                                "\u001b[32m2026-03-26 01:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5fcfbccf-d31a-4b57-955a-42b16c057297 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 01:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5fcfbccf-d31a-4b57-955a-42b16c057297 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy custom network devices fact] ****************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Thursday 26 March 2026  01:59:38 +0000 (0:00:00.093)       0:00:00.093 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Copy fact file] **********************************************************",
                                "Thursday 26 March 2026  01:59:40 +0000 (0:00:01.381)       0:00:01.475 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Copy custom ceph devices facts] ******************************************",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Thursday 26 March 2026  01:59:41 +0000 (0:00:01.181)       0:00:02.657 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Thursday 26 March 2026  01:59:41 +0000 (0:00:00.110)       0:00:02.767 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Thursday 26 March 2026  01:59:41 +0000 (0:00:00.205)       0:00:02.973 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Thursday 26 March 2026  01:59:41 +0000 (0:00:00.238)       0:00:03.211 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Thursday 26 March 2026  01:59:41 +0000 (0:00:00.152)       0:00:03.363 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Thursday 26 March 2026  01:59:42 +0000 (0:00:00.418)       0:00:03.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Thursday 26 March 2026  01:59:42 +0000 (0:00:00.139)       0:00:03.921 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Thursday 26 March 2026  01:59:43 +0000 (0:00:01.063)       0:00:04.985 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Thursday 26 March 2026  01:59:43 +0000 (0:00:00.474)       0:00:05.459 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Thursday 26 March 2026  01:59:45 +0000 (0:00:01.094)       0:00:06.554 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Install required packages (RedHat)] **************************************",
                                "Thursday 26 March 2026  02:00:01 +0000 (0:00:16.562)       0:00:23.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (Debian)] **************************************",
                                "Thursday 26 March 2026  02:00:01 +0000 (0:00:00.105)       0:00:23.221 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Thursday 26 March 2026  02:00:09 +0000 (0:00:07.791)       0:00:31.013 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "Thursday 26 March 2026  02:00:10 +0000 (0:00:00.457)       0:00:31.470 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Thursday 26 March 2026  02:00:13 +0000 (0:00:03.462)       0:00:34.933 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:00:14 +0000 (0:00:01.500)       0:00:36.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:00:18 +0000 (0:00:03.653)       0:00:40.088 ******** ",
                                "=============================================================================== ",
                                "osism.commons.repository : Update package cache ------------------------ 16.56s",
                                "Install required packages (Debian) -------------------------------------- 7.79s",
                                "Gathers facts about hosts ----------------------------------------------- 3.65s",
                                "Copy fact files --------------------------------------------------------- 3.46s",
                                "osism.commons.repository : Force update of package cache ---------------- 1.50s",
                                "Create custom facts directory ------------------------------------------- 1.38s",
                                "Copy fact file ---------------------------------------------------------- 1.18s",
                                "osism.commons.repository : Copy ubuntu.sources file --------------------- 1.09s",
                                "osism.commons.repository : Copy 99osism apt configuration --------------- 1.06s",
                                "osism.commons.repository : Remove sources.list file --------------------- 0.47s",
                                "Create custom facts directory ------------------------------------------- 0.46s",
                                "osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.42s",
                                "osism.commons.repository : Set repositories to default ------------------ 0.24s",
                                "osism.commons.repository : Set repository_default fact to default value --- 0.21s",
                                "osism.commons.repository : Include distribution specific repository tasks --- 0.15s",
                                "osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.14s",
                                "osism.commons.repository : Gather variables for each operating system --- 0.11s",
                                "Install required packages (RedHat) -------------------------------------- 0.11s",
                                "+ osism apply bootstrap",
                                "\u001b[32m2026-03-26 02:00:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4680eed5-a920-43b6-870f-1050432d97e6 (bootstrap) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:00:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4680eed5-a920-43b6-870f-1050432d97e6 (bootstrap) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on state bootstrap] ************************************",
                                "",
                                "TASK [Group hosts based on state bootstrap] ************************************",
                                "Thursday 26 March 2026  02:00:36 +0000 (0:00:00.171)       0:00:00.171 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:00:36 +0000 (0:00:00.288)       0:00:00.459 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts (if using --limit)] ***************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:00:40 +0000 (0:00:03.598)       0:00:04.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 1] ********************************************",
                                "",
                                "TASK [osism.commons.hostname : Set hostname] ***********************************",
                                "Thursday 26 March 2026  02:00:40 +0000 (0:00:00.675)       0:00:04.734 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.hostname : Copy /etc/hostname] *****************************",
                                "Thursday 26 March 2026  02:00:41 +0000 (0:00:01.214)       0:00:05.949 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Include type specific tasks] ***********************",
                                "Thursday 26 March 2026  02:00:43 +0000 (0:00:01.195)       0:00:07.145 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Copy /etc/hosts file] ******************************",
                                "Thursday 26 March 2026  02:00:43 +0000 (0:00:00.304)       0:00:07.450 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Include distribution specific tasks] ***************",
                                "Thursday 26 March 2026  02:00:45 +0000 (0:00:02.118)       0:00:09.568 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Configure proxy parameters for apt] ****************",
                                "Thursday 26 March 2026  02:00:45 +0000 (0:00:00.305)       0:00:09.874 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Set system wide settings in environment file] ******",
                                "Thursday 26 March 2026  02:00:46 +0000 (0:00:00.996)       0:00:10.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Remove system wide settings in environment file] ***",
                                "Thursday 26 March 2026  02:00:47 +0000 (0:00:00.561)       0:00:11.433 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Thursday 26 March 2026  02:00:47 +0000 (0:00:00.457)       0:00:11.890 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Thursday 26 March 2026  02:00:48 +0000 (0:00:00.243)       0:00:12.134 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Thursday 26 March 2026  02:00:48 +0000 (0:00:00.303)       0:00:12.437 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Thursday 26 March 2026  02:00:48 +0000 (0:00:00.326)       0:00:12.764 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Thursday 26 March 2026  02:00:50 +0000 (0:00:01.380)       0:00:14.144 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Thursday 26 March 2026  02:00:50 +0000 (0:00:00.353)       0:00:14.497 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Thursday 26 March 2026  02:00:51 +0000 (0:00:00.559)       0:00:15.057 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Thursday 26 March 2026  02:00:51 +0000 (0:00:00.281)       0:00:15.339 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Thursday 26 March 2026  02:00:51 +0000 (0:00:00.557)       0:00:15.896 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Thursday 26 March 2026  02:00:53 +0000 (0:00:01.134)       0:00:17.030 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Thursday 26 March 2026  02:00:54 +0000 (0:00:01.025)       0:00:18.056 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Thursday 26 March 2026  02:00:54 +0000 (0:00:00.329)       0:00:18.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Thursday 26 March 2026  02:00:55 +0000 (0:00:01.274)       0:00:19.660 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Thursday 26 March 2026  02:00:55 +0000 (0:00:00.243)       0:00:19.903 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Thursday 26 March 2026  02:00:56 +0000 (0:00:00.225)       0:00:20.129 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Thursday 26 March 2026  02:00:56 +0000 (0:00:00.274)       0:00:20.403 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Thursday 26 March 2026  02:00:56 +0000 (0:00:00.345)       0:00:20.749 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Thursday 26 March 2026  02:00:57 +0000 (0:00:00.521)       0:00:21.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Thursday 26 March 2026  02:00:57 +0000 (0:00:00.289)       0:00:21.560 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Thursday 26 March 2026  02:00:58 +0000 (0:00:01.103)       0:00:22.663 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Thursday 26 March 2026  02:00:59 +0000 (0:00:00.571)       0:00:23.235 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Thursday 26 March 2026  02:01:00 +0000 (0:00:01.168)       0:00:24.404 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Gather variables for each operating system] *****",
                                "Thursday 26 March 2026  02:01:16 +0000 (0:00:16.032)       0:00:40.436 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****",
                                "Thursday 26 March 2026  02:01:16 +0000 (0:00:00.214)       0:00:40.651 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***",
                                "Thursday 26 March 2026  02:01:16 +0000 (0:00:00.239)       0:00:40.890 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include distribution specific install tasks] ****",
                                "Thursday 26 March 2026  02:01:17 +0000 (0:00:00.239)       0:00:41.130 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Install rsyslog package] ************************",
                                "Thursday 26 March 2026  02:01:17 +0000 (0:00:00.333)       0:00:41.463 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********",
                                "Thursday 26 March 2026  02:01:19 +0000 (0:00:01.723)       0:00:43.187 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Manage rsyslog service] *************************",
                                "Thursday 26 March 2026  02:01:20 +0000 (0:00:01.058)       0:00:44.246 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include fluentd tasks] **************************",
                                "Thursday 26 March 2026  02:01:21 +0000 (0:00:00.824)       0:00:45.070 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***",
                                "Thursday 26 March 2026  02:01:21 +0000 (0:00:00.347)       0:00:45.418 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include additional log server tasks] ************",
                                "Thursday 26 March 2026  02:01:22 +0000 (0:00:01.002)       0:00:46.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include logrotate tasks] ************************",
                                "Thursday 26 March 2026  02:01:22 +0000 (0:00:00.255)       0:00:46.675 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Ensure logrotate package is installed] **********",
                                "Thursday 26 March 2026  02:01:22 +0000 (0:00:00.330)       0:00:47.006 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************",
                                "Thursday 26 March 2026  02:01:24 +0000 (0:00:01.779)       0:00:48.786 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Install util-linux-extra package] ****************",
                                "Thursday 26 March 2026  02:01:25 +0000 (0:00:01.139)       0:00:49.925 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Sync hardware clock] *****************************",
                                "Thursday 26 March 2026  02:01:39 +0000 (0:00:13.497)       0:01:03.423 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.configfs : Start sys-kernel-config mount] ******************",
                                "Thursday 26 March 2026  02:01:40 +0000 (0:00:01.248)       0:01:04.671 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Gather variables for each operating system] *****",
                                "Thursday 26 March 2026  02:01:41 +0000 (0:00:00.945)       0:01:05.617 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***",
                                "Thursday 26 March 2026  02:01:41 +0000 (0:00:00.272)       0:01:05.889 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Include distribution specific package tasks] ****",
                                "Thursday 26 March 2026  02:01:42 +0000 (0:00:00.253)       0:01:06.143 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install needrestart package] ********************",
                                "Thursday 26 March 2026  02:01:42 +0000 (0:00:00.316)       0:01:06.459 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set needrestart mode] ***************************",
                                "Thursday 26 March 2026  02:01:44 +0000 (0:00:01.604)       0:01:08.063 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***",
                                "Thursday 26 March 2026  02:01:44 +0000 (0:00:00.565)       0:01:08.628 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Update package cache] ***************************",
                                "Thursday 26 March 2026  02:01:44 +0000 (0:00:00.241)       0:01:08.870 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download upgrade packages] **********************",
                                "Thursday 26 March 2026  02:01:45 +0000 (0:00:01.143)       0:01:10.013 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Upgrade packages] *******************************",
                                "Thursday 26 March 2026  02:01:47 +0000 (0:00:01.661)       0:01:11.674 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download required packages] *********************",
                                "Thursday 26 March 2026  02:01:50 +0000 (0:00:03.133)       0:01:14.808 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install required packages] **********************",
                                "Thursday 26 March 2026  02:02:35 +0000 (0:00:44.819)       0:01:59.627 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove useless packages from the cache] *********",
                                "Thursday 26 March 2026  02:03:50 +0000 (0:01:14.819)       0:03:14.446 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove dependencies that are no longer required] ***",
                                "Thursday 26 March 2026  02:03:52 +0000 (0:00:01.685)       0:03:16.132 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Include sysctl tasks] *****************************",
                                "Thursday 26 March 2026  02:04:07 +0000 (0:00:15.146)       0:03:31.279 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********",
                                "Thursday 26 March 2026  02:04:07 +0000 (0:00:00.462)       0:03:31.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************",
                                "Thursday 26 March 2026  02:04:08 +0000 (0:00:00.829)       0:03:32.571 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************",
                                "Thursday 26 March 2026  02:04:13 +0000 (0:00:04.830)       0:03:37.401 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************",
                                "Thursday 26 March 2026  02:04:13 +0000 (0:00:00.583)       0:03:37.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on network] *****************",
                                "Thursday 26 March 2026  02:04:14 +0000 (0:00:00.524)       0:03:38.509 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************",
                                "Thursday 26 March 2026  02:04:15 +0000 (0:00:00.605)       0:03:39.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "",
                                "TASK [osism.commons.limits : Include limits tasks] *****************************",
                                "Thursday 26 March 2026  02:04:15 +0000 (0:00:00.611)       0:03:39.727 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Populate service facts] *************************",
                                "Thursday 26 March 2026  02:04:16 +0000 (0:00:00.365)       0:03:40.092 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Check services] *********************************",
                                "Thursday 26 March 2026  02:04:22 +0000 (0:00:06.599)       0:03:46.692 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Start/enable required services] *****************",
                                "Thursday 26 March 2026  02:04:23 +0000 (0:00:00.338)       0:03:47.031 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Include distribution specific configure tasks] ******",
                                "Thursday 26 March 2026  02:04:24 +0000 (0:00:01.163)       0:03:48.195 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove update-motd package] *************************",
                                "Thursday 26 March 2026  02:04:24 +0000 (0:00:00.448)       0:03:48.643 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************",
                                "Thursday 26 March 2026  02:04:26 +0000 (0:00:01.429)       0:03:50.072 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Disable the dynamic motd-news service] **************",
                                "Thursday 26 March 2026  02:04:26 +0000 (0:00:00.660)       0:03:50.732 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********",
                                "Thursday 26 March 2026  02:04:27 +0000 (0:00:00.640)       0:03:51.373 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove pam_motd.so rule] ****************************",
                                "Thursday 26 March 2026  02:04:27 +0000 (0:00:00.601)       0:03:51.975 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489204.3152304, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489229.2050931, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489232.927024, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489223.8100085, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489221.6920967, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489235.4309287, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774489217.9835355, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy motd file] *************************************",
                                "Thursday 26 March 2026  02:04:28 +0000 (0:00:00.949)       0:03:52.925 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue file] ************************************",
                                "Thursday 26 March 2026  02:04:29 +0000 (0:00:01.067)       0:03:53.992 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue.net file] ********************************",
                                "Thursday 26 March 2026  02:04:31 +0000 (0:00:01.206)       0:03:55.199 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to print the motd] ********************",
                                "Thursday 26 March 2026  02:04:32 +0000 (0:00:01.160)       0:03:56.359 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to not print the motd] ****************",
                                "Thursday 26 March 2026  02:04:32 +0000 (0:00:00.362)       0:03:56.722 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Include distribution specific install tasks] ********",
                                "Thursday 26 March 2026  02:04:33 +0000 (0:00:00.797)       0:03:57.519 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Install rng package] ********************************",
                                "Thursday 26 March 2026  02:04:34 +0000 (0:00:00.518)       0:03:58.037 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Remove haveged package] *****************************",
                                "Thursday 26 March 2026  02:04:42 +0000 (0:00:08.264)       0:04:06.302 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Manage rng service] *********************************",
                                "Thursday 26 March 2026  02:04:43 +0000 (0:00:01.293)       0:04:07.595 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Gather variables for each operating system] ******",
                                "Thursday 26 March 2026  02:04:44 +0000 (0:00:01.132)       0:04:08.728 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***",
                                "Thursday 26 March 2026  02:04:45 +0000 (0:00:00.334)       0:04:09.063 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***",
                                "Thursday 26 March 2026  02:04:45 +0000 (0:00:00.329)       0:04:09.392 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Populate service facts] **************************",
                                "Thursday 26 March 2026  02:04:45 +0000 (0:00:00.344)       0:04:09.737 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include distribution specific timer tasks] *******",
                                "Thursday 26 March 2026  02:04:51 +0000 (0:00:05.484)       0:04:15.221 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Disable apt-daily timers] ************************",
                                "Thursday 26 March 2026  02:04:51 +0000 (0:00:00.456)       0:04:15.678 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include service tasks] ***************************",
                                "Thursday 26 March 2026  02:04:52 +0000 (0:00:00.432)       0:04:16.111 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup services] ********************************",
                                "Thursday 26 March 2026  02:04:52 +0000 (0:00:00.463)       0:04:16.574 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include packages tasks] **************************",
                                "Thursday 26 March 2026  02:04:52 +0000 (0:00:00.391)       0:04:16.965 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup installed packages] **********************",
                                "Thursday 26 March 2026  02:04:53 +0000 (0:00:00.463)       0:04:17.429 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloudinit package] ************************",
                                "Thursday 26 March 2026  02:05:28 +0000 (0:00:35.079)       0:04:52.509 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********",
                                "Thursday 26 March 2026  02:05:36 +0000 (0:00:08.100)       0:05:00.609 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove useless packages from the cache] **********",
                                "Thursday 26 March 2026  02:05:44 +0000 (0:00:08.292)       0:05:08.902 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***",
                                "Thursday 26 March 2026  02:05:46 +0000 (0:00:01.785)       0:05:10.688 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include cloudinit tasks] *************************",
                                "Thursday 26 March 2026  02:05:52 +0000 (0:00:06.100)       0:05:16.788 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******",
                                "Thursday 26 March 2026  02:05:53 +0000 (0:00:00.464)       0:05:17.253 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Install tzdata package] *************************",
                                "Thursday 26 March 2026  02:05:53 +0000 (0:00:00.750)       0:05:18.003 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Set timezone to UTC] ****************************",
                                "Thursday 26 March 2026  02:05:55 +0000 (0:00:01.731)       0:05:19.735 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Create /etc/adjtime file] ***********************",
                                "Thursday 26 March 2026  02:05:56 +0000 (0:00:00.814)       0:05:20.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************",
                                "Thursday 26 March 2026  02:05:56 +0000 (0:00:00.365)       0:05:20.915 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather variables for each operating system] ******",
                                "Thursday 26 March 2026  02:05:57 +0000 (0:00:00.500)       0:05:21.416 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_version variable to default value] ****",
                                "Thursday 26 March 2026  02:05:57 +0000 (0:00:00.403)       0:05:21.819 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_cli_version variable to default value] ***",
                                "Thursday 26 March 2026  02:05:58 +0000 (0:00:00.364)       0:05:22.183 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker version] ***********************",
                                "Thursday 26 March 2026  02:05:58 +0000 (0:00:00.356)       0:05:22.539 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker cli version] *******************",
                                "Thursday 26 March 2026  02:05:58 +0000 (0:00:00.331)       0:05:22.871 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include block storage tasks] *********************",
                                "Thursday 26 March 2026  02:05:59 +0000 (0:00:00.349)       0:05:23.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include zram storage tasks] **********************",
                                "Thursday 26 March 2026  02:05:59 +0000 (0:00:00.327)       0:05:23.548 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include docker install tasks] ********************",
                                "Thursday 26 March 2026  02:05:59 +0000 (0:00:00.306)       0:05:23.854 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove old architecture-dependent repository] ****",
                                "Thursday 26 March 2026  02:06:00 +0000 (0:00:00.433)       0:05:24.288 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather package facts] ****************************",
                                "Thursday 26 March 2026  02:06:01 +0000 (0:00:01.019)       0:05:25.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Check whether packages are installed that should not be installed] ***",
                                "Thursday 26 March 2026  02:06:04 +0000 (0:00:03.173)       0:05:28.482 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install apt-transport-https package] *************",
                                "Thursday 26 March 2026  02:06:05 +0000 (0:00:00.724)       0:05:29.207 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository gpg key] **************************",
                                "Thursday 26 March 2026  02:06:11 +0000 (0:00:06.691)       0:05:35.898 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository] **********************************",
                                "Thursday 26 March 2026  02:06:13 +0000 (0:00:01.151)       0:05:37.050 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Update package cache] ****************************",
                                "Thursday 26 March 2026  02:06:21 +0000 (0:00:08.358)       0:05:45.408 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker package version] **********************",
                                "Thursday 26 March 2026  02:06:24 +0000 (0:00:03.275)       0:05:48.684 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker-cli package version] ******************",
                                "Thursday 26 March 2026  02:06:26 +0000 (0:00:01.393)       0:05:50.077 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unlock containerd package] ***********************",
                                "Thursday 26 March 2026  02:06:27 +0000 (0:00:01.803)       0:05:51.881 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install containerd package] **********************",
                                "Thursday 26 March 2026  02:06:28 +0000 (0:00:00.661)       0:05:52.542 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Lock containerd package] *************************",
                                "Thursday 26 March 2026  02:06:37 +0000 (0:00:08.552)       0:06:01.095 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker-cli package] **********************",
                                "Thursday 26 March 2026  02:06:38 +0000 (0:00:01.018)       0:06:02.114 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker package] **************************",
                                "Thursday 26 March 2026  02:06:46 +0000 (0:00:08.642)       0:06:10.756 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unblock installation of python docker packages] ***",
                                "Thursday 26 March 2026  02:06:57 +0000 (0:00:10.659)       0:06:21.415 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package] ******************",
                                "Thursday 26 March 2026  02:06:58 +0000 (0:00:01.231)       0:06:22.647 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package from Debian Sid] ***",
                                "Thursday 26 March 2026  02:06:59 +0000 (0:00:00.659)       0:06:23.306 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***",
                                "Thursday 26 March 2026  02:07:03 +0000 (0:00:03.719)       0:06:27.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***",
                                "Thursday 26 March 2026  02:07:03 +0000 (0:00:00.547)       0:06:27.574 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***",
                                "Thursday 26 March 2026  02:07:04 +0000 (0:00:00.853)       0:06:28.427 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***",
                                "Thursday 26 March 2026  02:07:04 +0000 (0:00:00.579)       0:06:29.006 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install packages required by docker login] *******",
                                "Thursday 26 March 2026  02:07:05 +0000 (0:00:00.498)       0:06:29.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Ensure that some packages are not installed] *****",
                                "Thursday 26 March 2026  02:07:06 +0000 (0:00:00.579)       0:06:30.084 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include config tasks] ****************************",
                                "Thursday 26 March 2026  02:07:08 +0000 (0:00:02.001)       0:06:32.086 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create plugins directory] ************************",
                                "Thursday 26 March 2026  02:07:09 +0000 (0:00:01.015)       0:06:33.101 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create systemd overlay directory] ****************",
                                "Thursday 26 March 2026  02:07:09 +0000 (0:00:00.889)       0:06:33.991 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy systemd overlay file] ***********************",
                                "Thursday 26 March 2026  02:07:10 +0000 (0:00:00.846)       0:06:34.838 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***",
                                "Thursday 26 March 2026  02:07:12 +0000 (0:00:01.747)       0:06:36.585 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy limits configuration file] ******************",
                                "Thursday 26 March 2026  02:07:14 +0000 (0:00:01.485)       0:06:38.070 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy daemon.json configuration file] *************",
                                "Thursday 26 March 2026  02:07:15 +0000 (0:00:01.332)       0:06:39.403 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include service tasks] ***************************",
                                "Thursday 26 March 2026  02:07:16 +0000 (0:00:01.554)       0:06:40.958 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon] ***************************",
                                "Thursday 26 March 2026  02:07:18 +0000 (0:00:01.118)       0:06:42.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage service] **********************************",
                                "Thursday 26 March 2026  02:07:19 +0000 (0:00:01.324)       0:06:43.400 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage docker socket service] ********************",
                                "Thursday 26 March 2026  02:07:21 +0000 (0:00:02.024)       0:06:45.425 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage containerd service] ***********************",
                                "Thursday 26 March 2026  02:07:22 +0000 (0:00:01.150)       0:06:46.576 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include bootstrap tasks] *************************",
                                "Thursday 26 March 2026  02:07:23 +0000 (0:00:01.315)       0:06:47.891 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:24 +0000 (0:00:01.010)       0:06:48.902 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:24 +0000 (0:00:00.043)       0:06:48.946 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:24 +0000 (0:00:00.049)       0:06:48.995 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:25 +0000 (0:00:00.043)       0:06:49.039 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:25 +0000 (0:00:00.048)       0:06:49.088 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:25 +0000 (0:00:00.048)       0:06:49.136 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Thursday 26 March 2026  02:07:25 +0000 (0:00:00.040)       0:06:49.177 ******** ",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Thursday 26 March 2026  02:07:25 +0000 (0:00:00.042)       0:06:49.220 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************",
                                "Thursday 26 March 2026  02:07:26 +0000 (0:00:01.242)       0:06:50.463 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********",
                                "Thursday 26 March 2026  02:07:28 +0000 (0:00:01.569)       0:06:52.032 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Restart docker service] ***************",
                                "Thursday 26 March 2026  02:07:29 +0000 (0:00:01.262)       0:06:53.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****",
                                "Thursday 26 March 2026  02:07:31 +0000 (0:00:02.423)       0:06:55.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add user to docker group] ************************",
                                "Thursday 26 March 2026  02:07:31 +0000 (0:00:00.106)       0:06:55.824 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Log into private registry and force re-authorization] ***",
                                "Thursday 26 March 2026  02:07:32 +0000 (0:00:01.117)       0:06:56.942 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include facts tasks] *****************************",
                                "Thursday 26 March 2026  02:07:33 +0000 (0:00:00.632)       0:06:57.574 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create facts directory] **************************",
                                "Thursday 26 March 2026  02:07:34 +0000 (0:00:01.197)       0:06:58.772 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy docker fact files] **************************",
                                "Thursday 26 March 2026  02:07:35 +0000 (0:00:00.853)       0:06:59.625 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : This install type is not supported] *******",
                                "Thursday 26 March 2026  02:07:38 +0000 (0:00:02.520)       0:07:02.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Include distribution specific install tasks] ***",
                                "Thursday 26 March 2026  02:07:38 +0000 (0:00:00.720)       0:07:02.866 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***",
                                "Thursday 26 March 2026  02:07:39 +0000 (0:00:00.835)       0:07:03.701 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******",
                                "Thursday 26 March 2026  02:07:40 +0000 (0:00:00.850)       0:07:04.552 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose binary] *************",
                                "Thursday 26 March 2026  02:07:41 +0000 (0:00:01.104)       0:07:05.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Uninstall docker-compose package] *********",
                                "Thursday 26 March 2026  02:07:42 +0000 (0:00:00.544)       0:07:06.201 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose script] ***************",
                                "Thursday 26 March 2026  02:07:43 +0000 (0:00:01.498)       0:07:07.700 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****",
                                "Thursday 26 March 2026  02:07:44 +0000 (0:00:00.616)       0:07:08.316 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********",
                                "Thursday 26 March 2026  02:07:52 +0000 (0:00:07.978)       0:07:16.295 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Enable osism.target] **********************",
                                "Thursday 26 March 2026  02:07:53 +0000 (0:00:01.620)       0:07:17.916 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****",
                                "Thursday 26 March 2026  02:07:55 +0000 (0:00:01.641)       0:07:19.557 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  02:07:57 +0000 (0:00:01.627)       0:07:21.185 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Thursday 26 March 2026  02:07:58 +0000 (0:00:00.872)       0:07:22.057 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check minimum and maximum number of servers] *****",
                                "Thursday 26 March 2026  02:07:59 +0000 (0:00:01.129)       0:07:23.186 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Gather variables for each operating system] ******",
                                "Thursday 26 March 2026  02:07:59 +0000 (0:00:00.559)       0:07:23.746 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***",
                                "Thursday 26 March 2026  02:08:00 +0000 (0:00:00.580)       0:07:24.326 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_key_file variable to default value] ***",
                                "Thursday 26 March 2026  02:08:01 +0000 (0:00:00.796)       0:07:25.123 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Populate service facts] **************************",
                                "Thursday 26 March 2026  02:08:01 +0000 (0:00:00.570)       0:07:25.693 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage timesyncd service] ************************",
                                "Thursday 26 March 2026  02:08:07 +0000 (0:00:05.626)       0:07:31.319 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Include distribution specific install tasks] *****",
                                "Thursday 26 March 2026  02:08:07 +0000 (0:00:00.601)       0:07:31.921 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Install package] *********************************",
                                "Thursday 26 March 2026  02:08:09 +0000 (0:00:01.100)       0:07:33.021 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage chrony service] ***************************",
                                "Thursday 26 March 2026  02:08:10 +0000 (0:00:01.951)       0:07:34.973 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check if configuration file exists] **************",
                                "Thursday 26 March 2026  02:08:12 +0000 (0:00:01.256)       0:07:36.229 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Copy configuration file] *************************",
                                "Thursday 26 March 2026  02:08:13 +0000 (0:00:00.875)       0:07:37.105 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Include distribution specific install tasks] ******",
                                "Thursday 26 March 2026  02:08:15 +0000 (0:00:02.118)       0:07:39.224 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Install lldpd package] ****************************",
                                "Thursday 26 March 2026  02:08:16 +0000 (0:00:00.919)       0:07:40.143 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Manage lldpd service] *****************************",
                                "Thursday 26 March 2026  02:08:25 +0000 (0:00:09.433)       0:07:49.577 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********",
                                "Thursday 26 March 2026  02:08:27 +0000 (0:00:01.978)       0:07:51.555 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************",
                                "Thursday 26 March 2026  02:08:28 +0000 (0:00:01.217)       0:07:52.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap role part 2] *********************************************",
                                "",
                                "TASK [Include hardening role] **************************************************",
                                "Thursday 26 March 2026  02:08:30 +0000 (0:00:01.322)       0:07:54.096 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 3] ********************************************",
                                "",
                                "TASK [osism.services.journald : Copy configuration file] ***********************",
                                "Thursday 26 March 2026  02:08:30 +0000 (0:00:00.768)       0:07:54.865 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.journald : Manage journald service] ***********************",
                                "Thursday 26 March 2026  02:08:32 +0000 (0:00:01.260)       0:07:56.126 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include auditd role] *****************************************************",
                                "Thursday 26 March 2026  02:08:33 +0000 (0:00:01.433)       0:07:57.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include smartd role] *****************************************************",
                                "Thursday 26 March 2026  02:08:34 +0000 (0:00:00.536)       0:07:58.095 ******** ",
                                "\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Include distribution specific install tasks] *****",
                                "Thursday 26 March 2026  02:08:35 +0000 (0:00:01.101)       0:07:59.197 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Install smartmontools package] *******************",
                                "Thursday 26 March 2026  02:08:36 +0000 (0:00:00.910)       0:08:00.108 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Create /var/log/smartd directory] ****************",
                                "Thursday 26 March 2026  02:08:44 +0000 (0:00:08.733)       0:08:08.841 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Copy smartmontools configuration file] ***********",
                                "Thursday 26 March 2026  02:08:45 +0000 (0:00:00.879)       0:08:09.721 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Manage smartd service] ***************************",
                                "Thursday 26 March 2026  02:08:47 +0000 (0:00:01.345)       0:08:11.067 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.journald : Restart journald service] ***********",
                                "Thursday 26 March 2026  02:08:49 +0000 (0:00:02.143)       0:08:13.211 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************",
                                "Thursday 26 March 2026  02:08:50 +0000 (0:00:01.251)       0:08:14.462 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Set state bootstrap] *****************************************************",
                                "",
                                "TASK [Set osism.bootstrap.status fact] *****************************************",
                                "Thursday 26 March 2026  02:08:51 +0000 (0:00:01.174)       0:08:15.636 ******** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  02:08:52 +0000 (0:00:00.861)       0:08:16.497 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Thursday 26 March 2026  02:08:53 +0000 (0:00:01.080)       0:08:17.578 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set osism.bootstrap.timestamp fact] **************************************",
                                "Thursday 26 March 2026  02:08:54 +0000 (0:00:01.166)       0:08:18.745 ******** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  02:08:55 +0000 (0:00:01.081)       0:08:19.826 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Thursday 26 March 2026  02:08:56 +0000 (0:00:00.897)       0:08:20.723 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:08:57 +0000 (0:00:01.223)       0:08:21.947 ******** ",
                                "=============================================================================== ",
                                "osism.commons.packages : Install required packages --------------------- 74.82s",
                                "osism.commons.packages : Download required packages -------------------- 44.82s",
                                "osism.commons.cleanup : Cleanup installed packages --------------------- 35.08s",
                                "osism.commons.repository : Update package cache ------------------------ 16.03s",
                                "osism.commons.packages : Remove dependencies that are no longer required -- 15.15s",
                                "osism.commons.systohc : Install util-linux-extra package --------------- 13.50s",
                                "osism.services.docker : Install docker package ------------------------- 10.66s",
                                "osism.services.lldpd : Install lldpd package ---------------------------- 9.43s",
                                "osism.services.smartd : Install smartmontools package ------------------- 8.73s",
                                "osism.services.docker : Install docker-cli package ---------------------- 8.64s",
                                "osism.services.docker : Install containerd package ---------------------- 8.55s",
                                "osism.services.docker : Add repository ---------------------------------- 8.36s",
                                "osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.29s",
                                "osism.services.rng : Install rng package -------------------------------- 8.26s",
                                "osism.commons.cleanup : Remove cloudinit package ------------------------ 8.10s",
                                "osism.commons.docker_compose : Install docker-compose-plugin package ---- 7.98s",
                                "osism.services.docker : Install apt-transport-https package ------------- 6.69s",
                                "osism.commons.services : Populate service facts ------------------------- 6.60s",
                                "osism.commons.cleanup : Remove dependencies that are no longer required --- 6.10s",
                                "osism.services.chrony : Populate service facts -------------------------- 5.63s",
                                "+ osism apply fail2ban",
                                "\u001b[32m2026-03-26 02:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fcfdf61c-4c97-405f-bf69-c9730546e312 (fail2ban) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fcfdf61c-4c97-405f-bf69-c9730546e312 (fail2ban) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role fail2ban] *****************************************************",
                                "",
                                "TASK [osism.services.fail2ban : Include distribution specific install tasks] ***",
                                "Thursday 26 March 2026  02:09:17 +0000 (0:00:00.287)       0:00:00.287 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Install fail2ban package] **********************",
                                "Thursday 26 March 2026  02:09:18 +0000 (0:00:01.216)       0:00:01.504 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Copy configuration files] **********************",
                                "Thursday 26 March 2026  02:09:29 +0000 (0:00:11.356)       0:00:12.860 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Manage fail2ban service] ***********************",
                                "Thursday 26 March 2026  02:09:31 +0000 (0:00:01.453)       0:00:14.313 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Reload fail2ban configuration] *****************",
                                "Thursday 26 March 2026  02:09:32 +0000 (0:00:01.479)       0:00:15.793 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:09:34 +0000 (0:00:01.671)       0:00:17.465 ******** ",
                                "=============================================================================== ",
                                "osism.services.fail2ban : Install fail2ban package --------------------- 11.36s",
                                "osism.services.fail2ban : Reload fail2ban configuration ----------------- 1.67s",
                                "osism.services.fail2ban : Manage fail2ban service ----------------------- 1.48s",
                                "osism.services.fail2ban : Copy configuration files ---------------------- 1.45s",
                                "osism.services.fail2ban : Include distribution specific install tasks --- 1.22s",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ osism apply network",
                                "\u001b[32m2026-03-26 02:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d321b30-b3f8-4748-8c86-c02309be4909 (network) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d321b30-b3f8-4748-8c86-c02309be4909 (network) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role network] ******************************************************",
                                "",
                                "TASK [osism.commons.network : Gather variables for each operating system] ******",
                                "Thursday 26 March 2026  02:09:52 +0000 (0:00:00.294)       0:00:00.294 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include type specific tasks] *********************",
                                "Thursday 26 March 2026  02:09:52 +0000 (0:00:00.779)       0:00:01.073 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install required packages] ***********************",
                                "Thursday 26 March 2026  02:09:54 +0000 (0:00:01.368)       0:00:02.442 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove ifupdown package] *************************",
                                "Thursday 26 March 2026  02:09:56 +0000 (0:00:02.009)       0:00:04.452 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create required directories] *********************",
                                "Thursday 26 March 2026  02:09:58 +0000 (0:00:01.785)       0:00:06.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Prepare netplan configuration template] **********",
                                "Thursday 26 March 2026  02:09:59 +0000 (0:00:00.991)       0:00:07.229 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy netplan configuration] **********************",
                                "Thursday 26 March 2026  02:10:03 +0000 (0:00:03.990)       0:00:11.220 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove netplan configuration template] ***********",
                                "Thursday 26 March 2026  02:10:04 +0000 (0:00:01.673)       0:00:12.894 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Check if path for interface file exists] *********",
                                "Thursday 26 March 2026  02:10:06 +0000 (0:00:01.953)       0:00:14.847 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy interfaces file] ****************************",
                                "Thursday 26 March 2026  02:10:07 +0000 (0:00:01.259)       0:00:16.107 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install package networkd-dispatcher] *************",
                                "Thursday 26 March 2026  02:10:08 +0000 (0:00:00.764)       0:00:16.871 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy dispatcher scripts] *************************",
                                "Thursday 26 March 2026  02:10:10 +0000 (0:00:02.186)       0:00:19.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Manage service networkd-dispatcher] **************",
                                "Thursday 26 March 2026  02:10:11 +0000 (0:00:01.044)       0:00:20.102 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include cleanup tasks] ***************************",
                                "Thursday 26 March 2026  02:10:13 +0000 (0:00:01.758)       0:00:21.860 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Thursday 26 March 2026  02:10:15 +0000 (0:00:01.358)       0:00:23.218 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Set network_configured_files fact] ***************",
                                "Thursday 26 March 2026  02:10:16 +0000 (0:00:01.181)       0:00:24.400 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Thursday 26 March 2026  02:10:16 +0000 (0:00:00.713)       0:00:25.113 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include dummy interfaces] ************************",
                                "Thursday 26 March 2026  02:10:18 +0000 (0:00:01.255)       0:00:26.368 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include vxlan interfaces] ************************",
                                "Thursday 26 March 2026  02:10:18 +0000 (0:00:00.679)       0:00:27.048 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-manager, testbed-node-3, testbed-node-2, testbed-node-4, testbed-node-1, testbed-node-0, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd netdev files] ************",
                                "Thursday 26 March 2026  02:10:25 +0000 (0:00:06.448)       0:00:33.497 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd network files] ***********",
                                "Thursday 26 March 2026  02:10:32 +0000 (0:00:06.877)       0:00:40.374 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include networkd cleanup tasks] ******************",
                                "Thursday 26 March 2026  02:10:38 +0000 (0:00:06.228)       0:00:46.602 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Thursday 26 March 2026  02:10:39 +0000 (0:00:01.442)       0:00:48.045 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Thursday 26 March 2026  02:10:41 +0000 (0:00:01.441)       0:00:49.487 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************",
                                "Thursday 26 March 2026  02:10:44 +0000 (0:00:02.662)       0:00:52.150 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********",
                                "Thursday 26 March 2026  02:10:44 +0000 (0:00:00.724)       0:00:52.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:10:45 +0000 (0:00:00.786)       0:00:53.661 ******** ",
                                "=============================================================================== ",
                                "osism.commons.network : Create systemd networkd netdev files ------------ 6.88s",
                                "osism.commons.network : Include vxlan interfaces ------------------------ 6.45s",
                                "osism.commons.network : Create systemd networkd network files ----------- 6.23s",
                                "osism.commons.network : Prepare netplan configuration template ---------- 3.99s",
                                "osism.commons.network : Remove unused configuration files --------------- 2.66s",
                                "osism.commons.network : Install package networkd-dispatcher ------------- 2.19s",
                                "osism.commons.network : Install required packages ----------------------- 2.01s",
                                "osism.commons.network : Remove netplan configuration template ----------- 1.95s",
                                "osism.commons.network : Remove ifupdown package ------------------------- 1.79s",
                                "osism.commons.network : Manage service networkd-dispatcher -------------- 1.76s",
                                "osism.commons.network : Copy netplan configuration ---------------------- 1.67s",
                                "osism.commons.network : Include networkd cleanup tasks ------------------ 1.44s",
                                "osism.commons.network : List existing configuration files --------------- 1.44s",
                                "osism.commons.network : Include type specific tasks --------------------- 1.37s",
                                "osism.commons.network : Include cleanup tasks --------------------------- 1.36s",
                                "osism.commons.network : Check if path for interface file exists --------- 1.26s",
                                "osism.commons.network : Remove unused configuration files --------------- 1.26s",
                                "osism.commons.network : List existing configuration files --------------- 1.18s",
                                "osism.commons.network : Copy dispatcher scripts ------------------------- 1.04s",
                                "osism.commons.network : Create required directories --------------------- 0.99s",
                                "+ osism apply wireguard",
                                "\u001b[32m2026-03-26 02:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d2213621-6efb-4be0-b77b-16b9f0ccce98 (wireguard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d2213621-6efb-4be0-b77b-16b9f0ccce98 (wireguard) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role wireguard] ****************************************************",
                                "",
                                "TASK [osism.services.wireguard : Install iptables package] *********************",
                                "Thursday 26 March 2026  02:11:03 +0000 (0:00:00.253)       0:00:00.253 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Install wireguard package] ********************",
                                "Thursday 26 March 2026  02:11:05 +0000 (0:00:01.799)       0:00:02.053 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create public and private key - server] *******",
                                "Thursday 26 March 2026  02:11:12 +0000 (0:00:07.246)       0:00:09.300 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create preshared key] *************************",
                                "Thursday 26 March 2026  02:11:12 +0000 (0:00:00.571)       0:00:09.872 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get preshared key] ****************************",
                                "Thursday 26 March 2026  02:11:13 +0000 (0:00:00.497)       0:00:10.370 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get public key - server] **********************",
                                "Thursday 26 March 2026  02:11:14 +0000 (0:00:00.790)       0:00:11.160 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get private key - server] *********************",
                                "Thursday 26 March 2026  02:11:14 +0000 (0:00:00.423)       0:00:11.583 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy wg0.conf configuration file] *************",
                                "Thursday 26 March 2026  02:11:15 +0000 (0:00:00.423)       0:00:12.007 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy client configuration files] **************",
                                "Thursday 26 March 2026  02:11:16 +0000 (0:00:01.252)       0:00:13.259 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********",
                                "Thursday 26 March 2026  02:11:17 +0000 (0:00:00.946)       0:00:14.206 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************",
                                "Thursday 26 March 2026  02:11:19 +0000 (0:00:01.789)       0:00:15.996 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:11:20 +0000 (0:00:00.949)       0:00:16.946 ******** ",
                                "=============================================================================== ",
                                "osism.services.wireguard : Install wireguard package -------------------- 7.25s",
                                "osism.services.wireguard : Install iptables package --------------------- 1.80s",
                                "osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.79s",
                                "osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.25s",
                                "osism.services.wireguard : Restart wg0 service -------------------------- 0.95s",
                                "osism.services.wireguard : Copy client configuration files -------------- 0.95s",
                                "osism.services.wireguard : Get preshared key ---------------------------- 0.79s",
                                "osism.services.wireguard : Create public and private key - server ------- 0.57s",
                                "osism.services.wireguard : Create preshared key ------------------------- 0.50s",
                                "osism.services.wireguard : Get private key - server --------------------- 0.42s",
                                "osism.services.wireguard : Get public key - server ---------------------- 0.42s",
                                "+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0",
                                "100    14  100    14    0     0    168      0 --:--:-- --:--:-- --:--:--   170",
                                "+ osism apply --environment custom workarounds",
                                "\u001b[32m2026-03-26 02:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m",
                                "\u001b[32m2026-03-26 02:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 02ff3201-de62-47b0-bff2-96101897c72a (workarounds) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 02ff3201-de62-47b0-bff2-96101897c72a (workarounds) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on virtualization_role] ********************************",
                                "Thursday 26 March 2026  02:11:37 +0000 (0:00:00.131)       0:00:00.131 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on the manager node] *************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Thursday 26 March 2026  02:11:38 +0000 (0:00:00.840)       0:00:00.972 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on all other nodes] **************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Thursday 26 March 2026  02:11:41 +0000 (0:00:02.657)       0:00:03.629 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add custom CA certificates to non-manager nodes] *************************",
                                "",
                                "TASK [Copy custom CA certificates] *********************************************",
                                "Thursday 26 March 2026  02:11:43 +0000 (0:00:01.868)       0:00:05.497 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "",
                                "TASK [Run update-ca-certificates] **********************************************",
                                "Thursday 26 March 2026  02:11:44 +0000 (0:00:01.573)       0:00:07.071 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Run update-ca-trust] *****************************************************",
                                "Thursday 26 March 2026  02:11:48 +0000 (0:00:03.701)       0:00:10.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add a workaround service] ************************************************",
                                "",
                                "TASK [Copy workarounds.sh scripts] *********************************************",
                                "Thursday 26 March 2026  02:11:49 +0000 (0:00:00.802)       0:00:11.575 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Copy workarounds systemd unit file] **************************************",
                                "Thursday 26 March 2026  02:11:50 +0000 (0:00:01.725)       0:00:13.300 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Reload systemd daemon] ***************************************************",
                                "Thursday 26 March 2026  02:11:52 +0000 (0:00:01.832)       0:00:15.133 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable workarounds.service (Debian)] *************************************",
                                "Thursday 26 March 2026  02:11:54 +0000 (0:00:01.766)       0:00:16.900 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable and start workarounds.service (RedHat)] ***************************",
                                "Thursday 26 March 2026  02:11:56 +0000 (0:00:02.071)       0:00:18.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************",
                                "",
                                "TASK [Install python3-docker] **************************************************",
                                "Thursday 26 March 2026  02:11:57 +0000 (0:00:00.679)       0:00:19.650 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:11:59 +0000 (0:00:02.830)       0:00:22.481 ******** ",
                                "=============================================================================== ",
                                "Run update-ca-certificates ---------------------------------------------- 3.70s",
                                "Install python3-docker -------------------------------------------------- 2.83s",
                                "Apply netplan configuration --------------------------------------------- 2.66s",
                                "Enable workarounds.service (Debian) ------------------------------------- 2.07s",
                                "Apply netplan configuration --------------------------------------------- 1.87s",
                                "Copy workarounds systemd unit file -------------------------------------- 1.83s",
                                "Reload systemd daemon --------------------------------------------------- 1.77s",
                                "Copy workarounds.sh scripts --------------------------------------------- 1.73s",
                                "Copy custom CA certificates --------------------------------------------- 1.57s",
                                "Group hosts based on virtualization_role -------------------------------- 0.84s",
                                "Run update-ca-trust ----------------------------------------------------- 0.80s",
                                "Enable and start workarounds.service (RedHat) --------------------------- 0.68s",
                                "+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-26 02:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 82c5765b-791d-4916-9564-ec8dce0bd1ed (reboot) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 82c5765b-791d-4916-9564-ec8dce0bd1ed (reboot) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:17 +0000 (0:00:00.208)       0:00:00.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:17 +0000 (0:00:00.118)       0:00:00.326 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:18 +0000 (0:00:00.949)       0:00:01.276 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:18 +0000 (0:00:00.120)       0:00:01.396 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:18 +0000 (0:00:00.113)       0:00:01.510 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:19 +0000 (0:00:00.679)       0:00:02.189 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:19 +0000 (0:00:00.117)       0:00:02.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:19 +0000 (0:00:00.220)       0:00:02.527 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:20 +0000 (0:00:00.675)       0:00:03.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:20 +0000 (0:00:00.117)       0:00:03.321 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:20 +0000 (0:00:00.119)       0:00:03.440 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:21 +0000 (0:00:00.672)       0:00:04.112 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:21 +0000 (0:00:00.133)       0:00:04.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:21 +0000 (0:00:00.119)       0:00:04.366 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:22 +0000 (0:00:00.695)       0:00:05.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Thursday 26 March 2026  02:12:22 +0000 (0:00:00.125)       0:00:05.187 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Thursday 26 March 2026  02:12:22 +0000 (0:00:00.122)       0:00:05.310 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Thursday 26 March 2026  02:12:23 +0000 (0:00:00.667)       0:00:05.977 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:12:23 +0000 (0:00:00.037)       0:00:06.015 ******** ",
                                "=============================================================================== ",
                                "Reboot system - do not wait for the reboot to complete ------------------ 4.34s",
                                "Exit playbook, if user did not mean to reboot systems ------------------- 0.81s",
                                "Reboot system - wait for the reboot to complete ------------------------- 0.65s",
                                "+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-26 02:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5436eb36-cc72-4e39-ad74-719ebae5ca2c (wait-for-connection) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5436eb36-cc72-4e39-ad74-719ebae5ca2c (wait-for-connection) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wait until remote systems are reachable] *********************************",
                                "",
                                "TASK [Wait until remote system is reachable] ***********************************",
                                "Thursday 26 March 2026  02:12:40 +0000 (0:00:00.242)       0:00:00.242 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:12:52 +0000 (0:00:11.540)       0:00:11.783 ******** ",
                                "=============================================================================== ",
                                "Wait until remote system is reachable ---------------------------------- 11.54s",
                                "+ osism apply hddtemp",
                                "\u001b[32m2026-03-26 02:13:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c3f3a815-b5d4-49d2-b3ca-aab97cf0a9af (hddtemp) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:13:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c3f3a815-b5d4-49d2-b3ca-aab97cf0a9af (hddtemp) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role hddtemp] ******************************************************",
                                "",
                                "TASK [osism.services.hddtemp : Gather variables for each operating system] *****",
                                "Thursday 26 March 2026  02:13:09 +0000 (0:00:00.268)       0:00:00.268 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific install tasks] ****",
                                "Thursday 26 March 2026  02:13:10 +0000 (0:00:00.792)       0:00:01.060 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Remove hddtemp package] *************************",
                                "Thursday 26 March 2026  02:13:12 +0000 (0:00:01.332)       0:00:02.393 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************",
                                "Thursday 26 March 2026  02:13:14 +0000 (0:00:02.024)       0:00:04.418 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Check if drivetemp module is available] *********",
                                "Thursday 26 March 2026  02:13:15 +0000 (0:00:01.239)       0:00:05.658 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************",
                                "Thursday 26 March 2026  02:13:16 +0000 (0:00:01.295)       0:00:06.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Install lm-sensors] *****************************",
                                "Thursday 26 March 2026  02:13:17 +0000 (0:00:00.935)       0:00:07.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific service tasks] ****",
                                "Thursday 26 March 2026  02:13:30 +0000 (0:00:12.465)       0:00:20.355 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Manage lm-sensors service] **********************",
                                "Thursday 26 March 2026  02:13:31 +0000 (0:00:01.506)       0:00:21.861 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:13:33 +0000 (0:00:01.969)       0:00:23.830 ******** ",
                                "=============================================================================== ",
                                "osism.services.hddtemp : Install lm-sensors ---------------------------- 12.47s",
                                "osism.services.hddtemp : Remove hddtemp package ------------------------- 2.02s",
                                "osism.services.hddtemp : Manage lm-sensors service ---------------------- 1.97s",
                                "osism.services.hddtemp : Include distribution specific service tasks ---- 1.51s",
                                "osism.services.hddtemp : Include distribution specific install tasks ---- 1.33s",
                                "osism.services.hddtemp : Check if drivetemp module is available --------- 1.30s",
                                "osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.24s",
                                "osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.94s",
                                "osism.services.hddtemp : Gather variables for each operating system ----- 0.79s",
                                "++ semver 9.5.0 7.1.1",
                                "+ [[ 1 -ge 0 ]]",
                                "+ sudo systemctl restart manager.service",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "ARA in ceph-ansible already disabled.",
                                "ARA in kolla-ansible already disabled.",
                                "ARA in osism-ansible already disabled.",
                                "ARA in osism-kubernetes already disabled.",
                                "+ osism apply gather-facts",
                                "\u001b[32m2026-03-26 02:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2b1a1c1b-3e66-41d1-9d10-d820120a0f50 (gather-facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2b1a1c1b-3e66-41d1-9d10-d820120a0f50 (gather-facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:15:12 +0000 (0:00:00.237)       0:00:00.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Thursday 26 March 2026  02:15:20 +0000 (0:00:08.322)       0:00:08.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:15:21 +0000 (0:00:00.598)       0:00:09.158 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 8.32s",
                                "Gather facts for all hosts ---------------------------------------------- 0.60s",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry",
                                "+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images",
                                "+ [[ false == \\t\\r\\u\\e ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T02:15:22.316793Z",
                            "start": "2026-03-26T01:50:16.148963Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001b",
                        "name": "Deploy manager + bootstrap nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "/opt/configuration/scripts/deploy-services.sh"
                            ],
                            "delta": "1:45:28.286456",
                            "end": "2026-03-26 04:00:51.006502",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 /opt/configuration/scripts/deploy-services.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001c-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 02:15:22.720046",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY SERVICES\n\n+ set -e\n+ echo\n+ echo '# DEPLOY SERVICES'\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sh -c /opt/configuration/scripts/pull-images.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n\n# PULL IMAGES\n+ echo\n+ echo '# PULL IMAGES'\n+ echo\n\n++ semver 9.5.0 7.0.0\n+ [[ 1 -ge 0 ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-03-26 02:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-03-26 02:15:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06102fcd-70e1-4004-8e7e-1d35d1447f80 (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:15:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06102fcd-70e1-4004-8e7e-1d35d1447f80 is running in background. No more output. Check ARA for logs.\u001b[0m\n+ sh -c /opt/configuration/scripts/deploy/001-helpers.sh\n\u001b[32m2026-03-26 02:15:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0e8cb748-e270-4d9d-9e6d-8d1250c8a56c (cgit) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:15:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0e8cb748-e270-4d9d-9e6d-8d1250c8a56c is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-26 02:16:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f22e88-54cd-4802-bc55-577a23115959 (dotfiles) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:16:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f22e88-54cd-4802-bc55-577a23115959 is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-26 02:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 39bcad9c-bdae-4e22-b89c-b6f057833c05 (homer) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 39bcad9c-bdae-4e22-b89c-b6f057833c05 is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-26 02:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4212e8d9-a5f3-4e44-a65a-eec554ca52dc (phpmyadmin) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4212e8d9-a5f3-4e44-a65a-eec554ca52dc is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-26 02:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b5a31f83-3df5-4041-9b65-a060944e7586 (sosreport) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b5a31f83-3df5-4041-9b65-a060944e7586 is running in background. No more output. Check ARA for logs.\u001b[0m\n+ sh -c /opt/configuration/scripts/deploy/500-kubernetes.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n++ semver 9.5.0 8.0.3\n+ [[ 1 -ge 0 ]]\n+ osism apply frr\n\u001b[32m2026-03-26 02:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d590088-fe2b-4236-a786-40d95fe4bd4a (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d590088-fe2b-4236-a786-40d95fe4bd4a (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nThursday 26 March 2026  02:16:58 +0000 (0:00:00.547)       0:00:00.547 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nThursday 26 March 2026  02:16:59 +0000 (0:00:00.632)       0:00:01.180 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nThursday 26 March 2026  02:17:04 +0000 (0:00:05.690)       0:00:06.871 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nThursday 26 March 2026  02:17:17 +0000 (0:00:13.118)       0:00:19.989 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nThursday 26 March 2026  02:17:19 +0000 (0:00:01.409)       0:00:21.398 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nThursday 26 March 2026  02:17:21 +0000 (0:00:01.636)       0:00:23.035 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nThursday 26 March 2026  02:17:22 +0000 (0:00:01.493)       0:00:24.528 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nThursday 26 March 2026  02:17:22 +0000 (0:00:00.158)       0:00:24.686 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nThursday 26 March 2026  02:17:22 +0000 (0:00:00.193)       0:00:24.879 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nThursday 26 March 2026  02:17:23 +0000 (0:00:00.979)       0:00:25.859 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nThursday 26 March 2026  02:17:26 +0000 (0:00:02.627)       0:00:28.487 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.frr : Restart frr service] *********************\nThursday 26 March 2026  02:17:28 +0000 (0:00:02.013)       0:00:30.500 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:17:30 +0000 (0:00:01.549)       0:00:32.050 ******** \n=============================================================================== \nosism.services.frr : Install frr package ------------------------------- 13.12s\nosism.services.frr : Pin frr package version ---------------------------- 5.69s\nosism.services.frr : Set sysctl parameters ------------------------------ 2.63s\nosism.services.frr : Manage frr service --------------------------------- 2.01s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 1.64s\nosism.services.frr : Restart frr service -------------------------------- 1.55s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 1.49s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.41s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.98s\nosism.services.frr : Include distribution specific install tasks -------- 0.63s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 0.19s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 0.16s\n+ osism apply kubernetes\n\u001b[32m2026-03-26 02:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5f21ae79-fb4c-403b-8e3d-454a1387a6c5 (kubernetes) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5f21ae79-fb4c-403b-8e3d-454a1387a6c5 (kubernetes) has been started and output is visible here.\u001b[0m\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nThursday 26 March 2026  02:17:38 +0000 (0:00:00.182)       0:00:00.182 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nThursday 26 March 2026  02:17:39 +0000 (0:00:00.753)       0:00:00.935 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nThursday 26 March 2026  02:17:40 +0000 (0:00:00.641)       0:00:01.577 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nThursday 26 March 2026  02:17:40 +0000 (0:00:00.818)       0:00:02.396 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nThursday 26 March 2026  02:17:42 +0000 (0:00:02.007)       0:00:04.403 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nThursday 26 March 2026  02:17:45 +0000 (0:00:02.155)       0:00:06.558 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nThursday 26 March 2026  02:17:46 +0000 (0:00:01.182)       0:00:07.741 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nThursday 26 March 2026  02:17:47 +0000 (0:00:00.898)       0:00:08.639 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nThursday 26 March 2026  02:17:47 +0000 (0:00:00.604)       0:00:09.244 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nThursday 26 March 2026  02:17:48 +0000 (0:00:00.709)       0:00:09.953 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nThursday 26 March 2026  02:17:49 +0000 (0:00:01.295)       0:00:11.249 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nThursday 26 March 2026  02:17:50 +0000 (0:00:00.855)       0:00:12.104 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nThursday 26 March 2026  02:17:55 +0000 (0:00:04.830)       0:00:16.935 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nThursday 26 March 2026  02:17:56 +0000 (0:00:01.074)       0:00:18.009 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nThursday 26 March 2026  02:17:57 +0000 (0:00:01.474)       0:00:19.483 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nThursday 26 March 2026  02:17:58 +0000 (0:00:00.776)       0:00:20.260 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nThursday 26 March 2026  02:17:59 +0000 (0:00:01.121)       0:00:21.381 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nThursday 26 March 2026  02:18:00 +0000 (0:00:00.675)       0:00:22.057 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nThursday 26 March 2026  02:18:01 +0000 (0:00:01.276)       0:00:23.334 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nThursday 26 March 2026  02:18:03 +0000 (0:00:01.765)       0:00:25.099 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nThursday 26 March 2026  02:18:05 +0000 (0:00:01.539)       0:00:26.639 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nThursday 26 March 2026  02:18:06 +0000 (0:00:01.156)       0:00:27.795 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nThursday 26 March 2026  02:18:07 +0000 (0:00:00.839)       0:00:28.635 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nThursday 26 March 2026  02:18:07 +0000 (0:00:00.419)       0:00:29.054 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nThursday 26 March 2026  02:18:08 +0000 (0:00:01.204)       0:00:30.259 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nThursday 26 March 2026  02:18:10 +0000 (0:00:01.634)       0:00:31.893 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nThursday 26 March 2026  02:18:11 +0000 (0:00:00.709)       0:00:32.603 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nThursday 26 March 2026  02:18:14 +0000 (0:00:02.900)       0:00:35.503 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nThursday 26 March 2026  02:18:14 +0000 (0:00:00.553)       0:00:36.056 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nThursday 26 March 2026  02:18:15 +0000 (0:00:00.796)       0:00:36.853 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nThursday 26 March 2026  02:18:16 +0000 (0:00:01.279)       0:00:38.133 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nThursday 26 March 2026  02:18:17 +0000 (0:00:00.802)       0:00:38.935 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nThursday 26 March 2026  02:18:17 +0000 (0:00:00.398)       0:00:39.334 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nThursday 26 March 2026  02:18:18 +0000 (0:00:01.123)       0:00:40.457 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nThursday 26 March 2026  02:18:22 +0000 (0:00:03.250)       0:00:43.707 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nThursday 26 March 2026  02:18:22 +0000 (0:00:00.530)       0:00:44.238 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nThursday 26 March 2026  02:19:16 +0000 (0:00:54.114)       0:01:38.353 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nThursday 26 March 2026  02:19:17 +0000 (0:00:00.342)       0:01:38.695 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nThursday 26 March 2026  02:19:18 +0000 (0:00:00.993)       0:01:39.689 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nThursday 26 March 2026  02:19:19 +0000 (0:00:01.187)       0:01:40.877 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nThursday 26 March 2026  02:19:46 +0000 (0:00:27.412)       0:02:08.290 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nThursday 26 March 2026  02:19:47 +0000 (0:00:00.649)       0:02:08.940 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nThursday 26 March 2026  02:19:48 +0000 (0:00:00.639)       0:02:09.580 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nThursday 26 March 2026  02:19:48 +0000 (0:00:00.667)       0:02:10.248 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nThursday 26 March 2026  02:19:49 +0000 (0:00:00.819)       0:02:11.067 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nThursday 26 March 2026  02:19:49 +0000 (0:00:00.332)       0:02:11.400 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nThursday 26 March 2026  02:19:50 +0000 (0:00:00.665)       0:02:12.066 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nThursday 26 March 2026  02:19:51 +0000 (0:00:00.672)       0:02:12.739 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nThursday 26 March 2026  02:19:52 +0000 (0:00:01.122)       0:02:13.861 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nThursday 26 March 2026  02:19:53 +0000 (0:00:00.881)       0:02:14.742 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nThursday 26 March 2026  02:19:53 +0000 (0:00:00.310)       0:02:15.053 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nThursday 26 March 2026  02:19:53 +0000 (0:00:00.344)       0:02:15.397 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nThursday 26 March 2026  02:19:54 +0000 (0:00:00.639)       0:02:16.037 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nThursday 26 March 2026  02:19:55 +0000 (0:00:00.907)       0:02:16.944 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nThursday 26 March 2026  02:19:58 +0000 (0:00:03.018)       0:02:19.963 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nThursday 26 March 2026  02:19:58 +0000 (0:00:00.350)       0:02:20.314 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nThursday 26 March 2026  02:19:59 +0000 (0:00:00.880)       0:02:21.195 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nThursday 26 March 2026  02:20:00 +0000 (0:00:00.323)       0:02:21.518 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nThursday 26 March 2026  02:20:00 +0000 (0:00:00.493)       0:02:22.012 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nThursday 26 March 2026  02:20:01 +0000 (0:00:00.563)       0:02:22.575 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nThursday 26 March 2026  02:20:01 +0000 (0:00:00.353)       0:02:22.929 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nThursday 26 March 2026  02:20:01 +0000 (0:00:00.325)       0:02:23.254 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nThursday 26 March 2026  02:20:02 +0000 (0:00:00.631)       0:02:23.886 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nThursday 26 March 2026  02:20:03 +0000 (0:00:01.410)       0:02:25.296 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nThursday 26 March 2026  02:20:05 +0000 (0:00:01.225)       0:02:26.522 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nThursday 26 March 2026  02:20:15 +0000 (0:00:10.243)       0:02:36.766 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nThursday 26 March 2026  02:20:16 +0000 (0:00:01.053)       0:02:37.820 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nThursday 26 March 2026  02:20:16 +0000 (0:00:00.471)       0:02:38.291 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nThursday 26 March 2026  02:20:17 +0000 (0:00:00.556)       0:02:38.848 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nThursday 26 March 2026  02:20:18 +0000 (0:00:00.897)       0:02:39.745 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nThursday 26 March 2026  02:20:18 +0000 (0:00:00.645)       0:02:40.390 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nThursday 26 March 2026  02:20:20 +0000 (0:00:01.629)       0:02:42.020 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nThursday 26 March 2026  02:20:21 +0000 (0:00:00.880)       0:02:42.901 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nThursday 26 March 2026  02:20:21 +0000 (0:00:00.494)       0:02:43.396 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nThursday 26 March 2026  02:20:22 +0000 (0:00:00.485)       0:02:43.881 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nThursday 26 March 2026  02:20:22 +0000 (0:00:00.393)       0:02:44.275 ******** \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nThursday 26 March 2026  02:20:23 +0000 (0:00:00.262)       0:02:44.537 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nThursday 26 March 2026  02:20:23 +0000 (0:00:00.867)       0:02:45.405 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nThursday 26 March 2026  02:20:25 +0000 (0:00:01.679)       0:02:47.085 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nThursday 26 March 2026  02:20:26 +0000 (0:00:00.838)       0:02:47.923 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nThursday 26 March 2026  02:20:26 +0000 (0:00:00.493)       0:02:48.416 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nThursday 26 March 2026  02:20:34 +0000 (0:00:07.985)       0:02:56.402 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nThursday 26 March 2026  02:20:47 +0000 (0:00:12.989)       0:03:09.392 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nThursday 26 March 2026  02:20:48 +0000 (0:00:00.817)       0:03:10.209 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nThursday 26 March 2026  02:20:49 +0000 (0:00:00.328)       0:03:10.538 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nThursday 26 March 2026  02:20:49 +0000 (0:00:00.326)       0:03:10.865 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nThursday 26 March 2026  02:20:50 +0000 (0:00:00.775)       0:03:11.640 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nThursday 26 March 2026  02:20:51 +0000 (0:00:00.915)       0:03:12.556 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nThursday 26 March 2026  02:20:52 +0000 (0:00:01.096)       0:03:13.653 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nThursday 26 March 2026  02:20:52 +0000 (0:00:00.134)       0:03:13.787 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nThursday 26 March 2026  02:20:53 +0000 (0:00:01.046)       0:03:14.834 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nThursday 26 March 2026  02:20:53 +0000 (0:00:00.142)       0:03:14.976 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nThursday 26 March 2026  02:20:53 +0000 (0:00:00.115)       0:03:15.092 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nThursday 26 March 2026  02:20:53 +0000 (0:00:00.131)       0:03:15.224 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nThursday 26 March 2026  02:20:53 +0000 (0:00:00.122)       0:03:15.346 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nThursday 26 March 2026  02:20:59 +0000 (0:00:05.567)       0:03:20.914 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nThursday 26 March 2026  02:21:41 +0000 (0:00:42.445)       0:04:03.359 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nThursday 26 March 2026  02:21:43 +0000 (0:00:01.296)       0:04:04.656 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nThursday 26 March 2026  02:21:44 +0000 (0:00:01.828)       0:04:06.484 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nThursday 26 March 2026  02:21:46 +0000 (0:00:01.154)       0:04:07.639 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nThursday 26 March 2026  02:21:46 +0000 (0:00:00.135)       0:04:07.774 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nThursday 26 March 2026  02:21:48 +0000 (0:00:01.862)       0:04:09.636 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nThursday 26 March 2026  02:21:48 +0000 (0:00:00.359)       0:04:09.996 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nThursday 26 March 2026  02:21:49 +0000 (0:00:00.850)       0:04:10.847 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nThursday 26 March 2026  02:21:49 +0000 (0:00:00.388)       0:04:11.235 ******** \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nThursday 26 March 2026  02:21:49 +0000 (0:00:00.237)       0:04:11.472 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nThursday 26 March 2026  02:21:55 +0000 (0:00:05.469)       0:04:16.942 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nThursday 26 March 2026  02:21:56 +0000 (0:00:00.807)       0:04:17.750 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nThursday 26 March 2026  02:22:04 +0000 (0:00:07.998)       0:04:25.748 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nThursday 26 March 2026  02:22:04 +0000 (0:00:00.578)       0:04:26.327 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:22:05 +0000 (0:00:00.706)       0:04:27.033 ******** \n=============================================================================== \nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 54.12s\nk3s_server_post : Wait for Cilium resources ---------------------------- 42.45s\nk3s_server : Enable and check K3s service ------------------------------ 27.41s\nkubectl : Install required packages ------------------------------------ 12.99s\nk3s_agent : Manage k3s service ----------------------------------------- 10.24s\nManage labels ----------------------------------------------------------- 8.00s\nkubectl : Add repository Debian ----------------------------------------- 7.99s\nk3s_server_post : Install Cilium ---------------------------------------- 5.57s\nk9s : Install k9s packages ---------------------------------------------- 5.47s\nk3s_download : Download k3s binary x64 ---------------------------------- 4.83s\nk3s_server : Detect Kubernetes version for label compatibility ---------- 3.25s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 3.02s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.90s\nk3s_prereq : Enable IPv6 forwarding ------------------------------------- 2.15s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.01s\nk3s_server_post : Test for BGP config resources ------------------------- 1.86s\nk3s_server_post : Copy BGP manifests to first master -------------------- 1.83s\nk3s_server : Validating arguments against arg spec 'main' - Setup k3s servers --- 1.77s\nkubectl : Install apt-transport-https package --------------------------- 1.68s\nk3s_server : Create custom resolv.conf for k3s -------------------------- 1.63s\n+ osism apply copy-kubeconfig\n\u001b[32m2026-03-26 02:22:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33d65d64-967a-44bc-a1e2-7e6dd84a9f9b (copy-kubeconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:22:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 33d65d64-967a-44bc-a1e2-7e6dd84a9f9b (copy-kubeconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy kubeconfig to the configuration repository] *************************\n\nTASK [Get kubeconfig file] *****************************************************\nThursday 26 March 2026  02:22:22 +0000 (0:00:00.158)       0:00:00.158 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nThursday 26 March 2026  02:22:23 +0000 (0:00:00.792)       0:00:00.950 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig file] ****************************\nThursday 26 March 2026  02:22:24 +0000 (0:00:01.278)       0:00:02.229 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:22:25 +0000 (0:00:00.497)       0:00:02.727 ******** \n=============================================================================== \nWrite kubeconfig file --------------------------------------------------- 1.28s\nGet kubeconfig file ----------------------------------------------------- 0.79s\nChange server address in the kubeconfig file ---------------------------- 0.50s\n+ sh -c /opt/configuration/scripts/deploy/200-infrastructure.sh\n\u001b[32m2026-03-26 02:22:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8310beb1-1a99-4254-99a6-2a442c4f14cd (openstackclient) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:22:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8310beb1-1a99-4254-99a6-2a442c4f14cd (openstackclient) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nThursday 26 March 2026  02:22:42 +0000 (0:00:00.237)       0:00:00.237 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nThursday 26 March 2026  02:22:43 +0000 (0:00:00.248)       0:00:00.485 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nThursday 26 March 2026  02:22:44 +0000 (0:00:01.298)       0:00:01.784 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nThursday 26 March 2026  02:22:45 +0000 (0:00:01.495)       0:00:03.279 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nThursday 26 March 2026  02:23:21 +0000 (0:00:35.575)       0:00:38.855 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nThursday 26 March 2026  02:23:22 +0000 (0:00:00.990)       0:00:39.845 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nThursday 26 March 2026  02:23:23 +0000 (0:00:00.734)       0:00:40.579 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nThursday 26 March 2026  02:23:24 +0000 (0:00:01.728)       0:00:42.308 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nThursday 26 March 2026  02:23:25 +0000 (0:00:00.772)       0:00:43.081 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nThursday 26 March 2026  02:23:26 +0000 (0:00:00.678)       0:00:43.760 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:23:26 +0000 (0:00:00.423)       0:00:44.183 ******** \n=============================================================================== \nosism.services.openstackclient : Manage openstackclient service -------- 35.58s\nosism.services.openstackclient : Restart openstackclient service -------- 1.73s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 1.50s\nosism.services.openstackclient : Create required directories ------------ 1.30s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 0.99s\nosism.services.openstackclient : Ensure that all containers are up ------ 0.77s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 0.73s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.68s\nosism.services.openstackclient : Copy bash completion script ------------ 0.42s\nosism.services.openstackclient : Include tasks -------------------------- 0.25s\n\u001b[32m2026-03-26 02:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a0a4ffc3-3e59-4deb-90a6-de603effbb9d (common) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a0a4ffc3-3e59-4deb-90a6-de603effbb9d (common) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  02:23:34 +0000 (0:00:00.291)       0:00:00.291 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nThursday 26 March 2026  02:23:35 +0000 (0:00:01.481)       0:00:01.773 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  02:23:38 +0000 (0:00:02.602)       0:00:04.375 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nThursday 26 March 2026  02:23:39 +0000 (0:00:01.386)       0:00:05.761 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:23:43 +0000 (0:00:03.494)       0:00:09.256 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nThursday 26 March 2026  02:23:44 +0000 (0:00:00.918)       0:00:10.174 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nThursday 26 March 2026  02:23:46 +0000 (0:00:02.013)       0:00:12.188 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nThursday 26 March 2026  02:23:46 +0000 (0:00:00.719)       0:00:12.907 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nThursday 26 March 2026  02:23:47 +0000 (0:00:00.943)       0:00:13.850 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nThursday 26 March 2026  02:23:51 +0000 (0:00:03.512)       0:00:17.363 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nThursday 26 March 2026  02:23:52 +0000 (0:00:00.993)       0:00:18.357 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nThursday 26 March 2026  02:23:53 +0000 (0:00:01.351)       0:00:19.709 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nThursday 26 March 2026  02:23:54 +0000 (0:00:00.865)       0:00:20.574 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nThursday 26 March 2026  02:23:55 +0000 (0:00:00.949)       0:00:21.524 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nThursday 26 March 2026  02:23:58 +0000 (0:00:02.586)       0:00:24.110 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nThursday 26 March 2026  02:24:00 +0000 (0:00:02.216)       0:00:26.326 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nThursday 26 March 2026  02:24:02 +0000 (0:00:02.051)       0:00:28.377 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nThursday 26 March 2026  02:24:03 +0000 (0:00:01.523)       0:00:29.901 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nThursday 26 March 2026  02:24:05 +0000 (0:00:01.957)       0:00:31.859 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [common : Check common containers] ****************************************\nThursday 26 March 2026  02:24:07 +0000 (0:00:01.692)       0:00:33.552 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Creating log volume] ********************************************\nThursday 26 March 2026  02:24:10 +0000 (0:00:02.613)       0:00:36.166 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Link kolla_logs volume to /var/log/kolla] ***********************\nThursday 26 March 2026  02:24:11 +0000 (0:00:01.411)       0:00:37.577 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:12 +0000 (0:00:01.071)       0:00:38.648 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:12 +0000 (0:00:00.065)       0:00:38.713 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:12 +0000 (0:00:00.066)       0:00:38.780 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:12 +0000 (0:00:00.088)       0:00:38.869 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:13 +0000 (0:00:00.252)       0:00:39.121 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:13 +0000 (0:00:00.067)       0:00:39.188 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  02:24:13 +0000 (0:00:00.059)       0:00:39.248 ******** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nThursday 26 March 2026  02:24:13 +0000 (0:00:00.090)       0:00:39.338 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nThursday 26 March 2026  02:24:53 +0000 (0:00:40.562)       0:01:19.900 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nThursday 26 March 2026  02:25:25 +0000 (0:00:31.651)       0:01:51.552 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nThursday 26 March 2026  02:25:27 +0000 (0:00:01.878)       0:01:53.430 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:25:35 +0000 (0:00:08.580)       0:02:02.011 ******** \n=============================================================================== \ncommon : Restart fluentd container ------------------------------------- 40.56s\ncommon : Restart kolla-toolbox container ------------------------------- 31.65s\ncommon : Restart cron container ----------------------------------------- 8.58s\ncommon : Copying over config.json files for services -------------------- 3.51s\nservice-cert-copy : common | Copying over extra CA certificates --------- 3.49s\ncommon : Check common containers ---------------------------------------- 2.61s\ncommon : Ensuring config directories exist ------------------------------ 2.60s\ncommon : Copying over fluentd.conf -------------------------------------- 2.59s\ncommon : Copying over cron logrotate config file ------------------------ 2.22s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.05s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 2.01s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 1.96s\ncommon : Initializing toolbox container using normal user --------------- 1.88s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 1.69s\ncommon : Ensuring config directories have correct owner and permission --- 1.52s\ncommon : include_tasks -------------------------------------------------- 1.48s\ncommon : Creating log volume -------------------------------------------- 1.41s\ncommon : include_tasks -------------------------------------------------- 1.39s\ncommon : Find custom fluentd filter config files ------------------------ 1.35s\ncommon : Link kolla_logs volume to /var/log/kolla ----------------------- 1.07s\n\u001b[32m2026-03-26 02:25:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69ee4675-db84-4d53-8664-3bcc753c27a0 (loadbalancer) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:25:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 69ee4675-db84-4d53-8664-3bcc753c27a0 (loadbalancer) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:25:43 +0000 (0:00:00.271)       0:00:00.271 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:25:43 +0000 (0:00:00.310)       0:00:00.582 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nThursday 26 March 2026  02:25:44 +0000 (0:00:00.455)       0:00:01.037 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nThursday 26 March 2026  02:25:44 +0000 (0:00:00.598)       0:00:01.635 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nThursday 26 March 2026  02:25:45 +0000 (0:00:00.619)       0:00:02.255 ******** \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nThursday 26 March 2026  02:25:45 +0000 (0:00:00.735)       0:00:02.991 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nThursday 26 March 2026  02:25:46 +0000 (0:00:00.601)       0:00:03.593 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nThursday 26 March 2026  02:25:48 +0000 (0:00:02.236)       0:00:05.829 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nThursday 26 March 2026  02:25:49 +0000 (0:00:00.742)       0:00:06.572 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nThursday 26 March 2026  02:25:50 +0000 (0:00:01.251)       0:00:07.823 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nThursday 26 March 2026  02:25:51 +0000 (0:00:00.553)       0:00:08.376 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nThursday 26 March 2026  02:25:53 +0000 (0:00:01.836)       0:00:10.213 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nThursday 26 March 2026  02:25:54 +0000 (0:00:00.910)       0:00:11.124 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nThursday 26 March 2026  02:25:55 +0000 (0:00:01.544)       0:00:12.669 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nThursday 26 March 2026  02:25:56 +0000 (0:00:00.931)       0:00:13.600 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nThursday 26 March 2026  02:25:57 +0000 (0:00:01.366)       0:00:14.967 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nThursday 26 March 2026  02:25:58 +0000 (0:00:00.605)       0:00:15.572 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nThursday 26 March 2026  02:26:01 +0000 (0:00:02.830)       0:00:18.402 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nThursday 26 March 2026  02:26:04 +0000 (0:00:03.217)       0:00:21.620 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nThursday 26 March 2026  02:26:06 +0000 (0:00:01.956)       0:00:23.576 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nThursday 26 March 2026  02:26:09 +0000 (0:00:03.037)       0:00:26.614 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nThursday 26 March 2026  02:26:10 +0000 (0:00:00.609)       0:00:27.223 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nThursday 26 March 2026  02:26:12 +0000 (0:00:02.106)       0:00:29.329 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nThursday 26 March 2026  02:26:14 +0000 (0:00:02.125)       0:00:31.455 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nThursday 26 March 2026  02:26:15 +0000 (0:00:01.462)       0:00:32.918 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nThursday 26 March 2026  02:26:17 +0000 (0:00:01.414)       0:00:34.332 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***\nThursday 26 March 2026  02:26:17 +0000 (0:00:00.556)       0:00:34.889 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:26:21 +0000 (0:00:03.351)       0:00:38.240 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***\nThursday 26 March 2026  02:26:21 +0000 (0:00:00.620)       0:00:38.861 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nThursday 26 March 2026  02:26:22 +0000 (0:00:00.800)       0:00:39.661 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:26:23 +0000 (0:00:01.018)       0:00:40.680 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nThursday 26 March 2026  02:26:24 +0000 (0:00:00.654)       0:00:41.334 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******\nThursday 26 March 2026  02:26:25 +0000 (0:00:00.822)       0:00:42.157 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:26:26 +0000 (0:00:01.113)       0:00:43.270 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****\nThursday 26 March 2026  02:26:26 +0000 (0:00:00.614)       0:00:43.885 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nThursday 26 March 2026  02:26:27 +0000 (0:00:00.804)       0:00:44.689 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nThursday 26 March 2026  02:26:29 +0000 (0:00:01.699)       0:00:46.389 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nThursday 26 March 2026  02:26:31 +0000 (0:00:01.762)       0:00:48.152 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Check loadbalancer containers] ****************************\nThursday 26 March 2026  02:26:31 +0000 (0:00:00.812)       0:00:48.964 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nThursday 26 March 2026  02:26:34 +0000 (0:00:02.431)       0:00:51.395 ******** \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nThursday 26 March 2026  02:26:35 +0000 (0:00:00.818)       0:00:52.213 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nThursday 26 March 2026  02:26:38 +0000 (0:00:03.286)       0:00:55.500 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nThursday 26 March 2026  02:26:39 +0000 (0:00:00.660)       0:00:56.161 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nThursday 26 March 2026  02:26:40 +0000 (0:00:01.127)       0:00:57.288 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nThursday 26 March 2026  02:26:41 +0000 (0:00:01.295)       0:00:58.584 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nThursday 26 March 2026  02:26:43 +0000 (0:00:02.068)       0:01:00.653 ******** \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nThursday 26 March 2026  02:26:44 +0000 (0:00:00.678)       0:01:01.331 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nThursday 26 March 2026  02:26:47 +0000 (0:00:03.402)       0:01:04.734 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nThursday 26 March 2026  02:26:48 +0000 (0:00:00.668)       0:01:05.403 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nThursday 26 March 2026  02:26:49 +0000 (0:00:00.828)       0:01:06.232 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nThursday 26 March 2026  02:26:50 +0000 (0:00:01.596)       0:01:07.828 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nThursday 26 March 2026  02:26:52 +0000 (0:00:02.053)       0:01:09.882 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nThursday 26 March 2026  02:26:53 +0000 (0:00:00.333)       0:01:10.215 ******** \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nThursday 26 March 2026  02:26:53 +0000 (0:00:00.738)       0:01:10.954 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nThursday 26 March 2026  02:26:56 +0000 (0:00:02.822)       0:01:13.776 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nThursday 26 March 2026  02:26:58 +0000 (0:00:01.575)       0:01:15.352 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nThursday 26 March 2026  02:26:59 +0000 (0:00:01.637)       0:01:16.990 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nThursday 26 March 2026  02:27:00 +0000 (0:00:00.440)       0:01:17.431 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nThursday 26 March 2026  02:27:01 +0000 (0:00:01.296)       0:01:18.727 ******** \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nThursday 26 March 2026  02:27:02 +0000 (0:00:01.012)       0:01:19.740 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nThursday 26 March 2026  02:27:06 +0000 (0:00:03.487)       0:01:23.227 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nThursday 26 March 2026  02:27:06 +0000 (0:00:00.759)       0:01:23.986 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nThursday 26 March 2026  02:27:08 +0000 (0:00:01.222)       0:01:25.209 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nThursday 26 March 2026  02:27:09 +0000 (0:00:01.266)       0:01:26.476 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nThursday 26 March 2026  02:27:11 +0000 (0:00:02.038)       0:01:28.515 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nThursday 26 March 2026  02:27:11 +0000 (0:00:00.313)       0:01:28.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nThursday 26 March 2026  02:27:12 +0000 (0:00:00.338)       0:01:29.167 ******** \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nThursday 26 March 2026  02:27:13 +0000 (0:00:00.976)       0:01:30.143 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nThursday 26 March 2026  02:27:16 +0000 (0:00:03.718)       0:01:33.861 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nThursday 26 March 2026  02:27:17 +0000 (0:00:01.078)       0:01:34.940 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nThursday 26 March 2026  02:27:19 +0000 (0:00:01.314)       0:01:36.254 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nThursday 26 March 2026  02:27:20 +0000 (0:00:01.332)       0:01:37.587 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nThursday 26 March 2026  02:27:22 +0000 (0:00:02.013)       0:01:39.600 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nThursday 26 March 2026  02:27:22 +0000 (0:00:00.354)       0:01:39.955 ******** \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nThursday 26 March 2026  02:27:24 +0000 (0:00:01.159)       0:01:41.114 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nThursday 26 March 2026  02:27:28 +0000 (0:00:04.520)       0:01:45.635 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nThursday 26 March 2026  02:27:31 +0000 (0:00:02.970)       0:01:48.605 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nThursday 26 March 2026  02:27:35 +0000 (0:00:04.351)       0:01:52.957 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nThursday 26 March 2026  02:27:37 +0000 (0:00:01.349)       0:01:54.307 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nThursday 26 March 2026  02:27:39 +0000 (0:00:02.134)       0:01:56.441 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nThursday 26 March 2026  02:27:39 +0000 (0:00:00.359)       0:01:56.800 ******** \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nThursday 26 March 2026  02:27:40 +0000 (0:00:01.123)       0:01:57.924 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nThursday 26 March 2026  02:27:43 +0000 (0:00:02.977)       0:02:00.901 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nThursday 26 March 2026  02:27:44 +0000 (0:00:00.400)       0:02:01.302 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nThursday 26 March 2026  02:27:45 +0000 (0:00:00.930)       0:02:02.232 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nThursday 26 March 2026  02:27:46 +0000 (0:00:01.294)       0:02:03.527 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nThursday 26 March 2026  02:27:48 +0000 (0:00:02.027)       0:02:05.555 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nThursday 26 March 2026  02:27:48 +0000 (0:00:00.342)       0:02:05.897 ******** \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nThursday 26 March 2026  02:27:50 +0000 (0:00:01.208)       0:02:07.106 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nThursday 26 March 2026  02:27:53 +0000 (0:00:03.313)       0:02:10.420 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nThursday 26 March 2026  02:27:54 +0000 (0:00:00.742)       0:02:11.162 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nThursday 26 March 2026  02:27:55 +0000 (0:00:01.167)       0:02:12.330 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nThursday 26 March 2026  02:27:56 +0000 (0:00:01.631)       0:02:13.962 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nThursday 26 March 2026  02:27:59 +0000 (0:00:02.205)       0:02:16.167 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nThursday 26 March 2026  02:27:59 +0000 (0:00:00.310)       0:02:16.478 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nThursday 26 March 2026  02:27:59 +0000 (0:00:00.346)       0:02:16.824 ******** \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nThursday 26 March 2026  02:28:01 +0000 (0:00:01.252)       0:02:18.077 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nThursday 26 March 2026  02:28:04 +0000 (0:00:03.224)       0:02:21.302 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nThursday 26 March 2026  02:28:04 +0000 (0:00:00.702)       0:02:22.004 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nThursday 26 March 2026  02:28:06 +0000 (0:00:01.183)       0:02:23.187 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nThursday 26 March 2026  02:28:07 +0000 (0:00:01.380)       0:02:24.568 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nThursday 26 March 2026  02:28:09 +0000 (0:00:02.118)       0:02:26.686 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nThursday 26 March 2026  02:28:09 +0000 (0:00:00.315)       0:02:27.002 ******** \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nThursday 26 March 2026  02:28:11 +0000 (0:00:01.413)       0:02:28.415 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nThursday 26 March 2026  02:28:14 +0000 (0:00:03.550)       0:02:31.966 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nThursday 26 March 2026  02:28:15 +0000 (0:00:00.753)       0:02:32.720 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nThursday 26 March 2026  02:28:16 +0000 (0:00:00.941)       0:02:33.661 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nThursday 26 March 2026  02:28:18 +0000 (0:00:01.687)       0:02:35.349 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nThursday 26 March 2026  02:28:20 +0000 (0:00:02.132)       0:02:37.481 ******** \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nThursday 26 March 2026  02:28:21 +0000 (0:00:01.120)       0:02:38.601 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nThursday 26 March 2026  02:28:25 +0000 (0:00:03.958)       0:02:42.560 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nThursday 26 March 2026  02:28:26 +0000 (0:00:01.091)       0:02:43.651 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nThursday 26 March 2026  02:28:27 +0000 (0:00:00.947)       0:02:44.599 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nThursday 26 March 2026  02:28:28 +0000 (0:00:01.299)       0:02:45.899 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nThursday 26 March 2026  02:28:31 +0000 (0:00:02.141)       0:02:48.041 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nThursday 26 March 2026  02:28:32 +0000 (0:00:01.387)       0:02:49.428 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nThursday 26 March 2026  02:28:35 +0000 (0:00:03.120)       0:02:52.548 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nThursday 26 March 2026  02:28:38 +0000 (0:00:02.467)       0:02:55.015 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nThursday 26 March 2026  02:28:40 +0000 (0:00:02.500)       0:02:57.516 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nThursday 26 March 2026  02:28:43 +0000 (0:00:03.062)       0:03:00.579 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nThursday 26 March 2026  02:28:45 +0000 (0:00:02.181)       0:03:02.760 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nThursday 26 March 2026  02:28:47 +0000 (0:00:01.460)       0:03:04.221 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nThursday 26 March 2026  02:28:47 +0000 (0:00:00.346)       0:03:04.568 ******** \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nThursday 26 March 2026  02:28:49 +0000 (0:00:01.452)       0:03:06.021 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nThursday 26 March 2026  02:28:50 +0000 (0:00:01.598)       0:03:07.619 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nThursday 26 March 2026  02:28:51 +0000 (0:00:00.405)       0:03:08.024 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nThursday 26 March 2026  02:28:51 +0000 (0:00:00.952)       0:03:08.977 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nThursday 26 March 2026  02:28:52 +0000 (0:00:00.443)       0:03:09.420 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nThursday 26 March 2026  02:28:53 +0000 (0:00:01.457)       0:03:10.878 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nThursday 26 March 2026  02:28:54 +0000 (0:00:00.351)       0:03:11.230 ******** \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nThursday 26 March 2026  02:28:55 +0000 (0:00:01.591)       0:03:12.822 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nThursday 26 March 2026  02:29:00 +0000 (0:00:04.359)       0:03:17.182 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nThursday 26 March 2026  02:29:01 +0000 (0:00:01.524)       0:03:18.706 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nThursday 26 March 2026  02:29:03 +0000 (0:00:02.029)       0:03:20.736 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nThursday 26 March 2026  02:29:05 +0000 (0:00:01.377)       0:03:22.113 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nThursday 26 March 2026  02:29:07 +0000 (0:00:02.167)       0:03:24.281 ******** \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nThursday 26 March 2026  02:29:08 +0000 (0:00:01.251)       0:03:25.533 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nThursday 26 March 2026  02:29:12 +0000 (0:00:03.611)       0:03:29.144 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nThursday 26 March 2026  02:29:12 +0000 (0:00:00.545)       0:03:29.690 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nThursday 26 March 2026  02:29:13 +0000 (0:00:00.799)       0:03:30.490 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nThursday 26 March 2026  02:29:15 +0000 (0:00:01.966)       0:03:32.457 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nThursday 26 March 2026  02:29:17 +0000 (0:00:01.930)       0:03:34.387 ******** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nThursday 26 March 2026  02:29:18 +0000 (0:00:01.583)       0:03:35.971 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nThursday 26 March 2026  02:29:23 +0000 (0:00:04.202)       0:03:40.173 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nThursday 26 March 2026  02:29:24 +0000 (0:00:01.431)       0:03:41.605 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nThursday 26 March 2026  02:29:25 +0000 (0:00:01.078)       0:03:42.683 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nThursday 26 March 2026  02:29:27 +0000 (0:00:01.507)       0:03:44.190 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nThursday 26 March 2026  02:29:29 +0000 (0:00:02.166)       0:03:46.356 ******** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nThursday 26 March 2026  02:29:31 +0000 (0:00:01.659)       0:03:48.016 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nThursday 26 March 2026  02:29:31 +0000 (0:00:00.881)       0:03:48.897 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nThursday 26 March 2026  02:29:35 +0000 (0:00:04.092)       0:03:52.990 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nThursday 26 March 2026  02:29:37 +0000 (0:00:01.469)       0:03:54.460 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  02:29:38 +0000 (0:00:01.527)       0:03:55.987 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  02:29:41 +0000 (0:00:02.483)       0:03:58.470 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nThursday 26 March 2026  02:29:44 +0000 (0:00:02.870)       0:04:01.341 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nThursday 26 March 2026  02:29:45 +0000 (0:00:01.117)       0:04:02.458 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nThursday 26 March 2026  02:29:46 +0000 (0:00:01.039)       0:04:03.498 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nThursday 26 March 2026  02:29:47 +0000 (0:00:01.291)       0:04:04.789 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  02:29:49 +0000 (0:00:01.606)       0:04:06.396 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  02:29:52 +0000 (0:00:02.774)       0:04:09.171 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nThursday 26 March 2026  02:29:54 +0000 (0:00:02.714)       0:04:11.885 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nThursday 26 March 2026  02:29:56 +0000 (0:00:01.265)       0:04:13.151 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nThursday 26 March 2026  02:29:57 +0000 (0:00:01.376)       0:04:14.527 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nThursday 26 March 2026  02:29:58 +0000 (0:00:01.422)       0:04:15.950 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  02:30:00 +0000 (0:00:01.896)       0:04:17.847 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  02:30:04 +0000 (0:00:03.266)       0:04:21.114 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nThursday 26 March 2026  02:30:07 +0000 (0:00:03.279)       0:04:24.394 ******** \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nThursday 26 March 2026  02:30:08 +0000 (0:00:01.483)       0:04:25.878 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nThursday 26 March 2026  02:30:12 +0000 (0:00:03.331)       0:04:29.209 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nThursday 26 March 2026  02:30:12 +0000 (0:00:00.741)       0:04:29.950 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nThursday 26 March 2026  02:30:13 +0000 (0:00:00.964)       0:04:30.915 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nThursday 26 March 2026  02:30:15 +0000 (0:00:01.802)       0:04:32.717 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nThursday 26 March 2026  02:30:17 +0000 (0:00:02.208)       0:04:34.926 ******** \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nThursday 26 March 2026  02:30:19 +0000 (0:00:01.424)       0:04:36.351 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nThursday 26 March 2026  02:30:24 +0000 (0:00:05.349)       0:04:41.700 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nThursday 26 March 2026  02:30:25 +0000 (0:00:01.069)       0:04:42.769 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nThursday 26 March 2026  02:30:26 +0000 (0:00:00.994)       0:04:43.764 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nThursday 26 March 2026  02:30:27 +0000 (0:00:00.451)       0:04:44.216 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nThursday 26 March 2026  02:30:28 +0000 (0:00:01.471)       0:04:45.688 ******** \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nThursday 26 March 2026  02:30:30 +0000 (0:00:01.846)       0:04:47.534 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nThursday 26 March 2026  02:30:34 +0000 (0:00:04.189)       0:04:51.723 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nThursday 26 March 2026  02:30:35 +0000 (0:00:00.854)       0:04:52.577 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nThursday 26 March 2026  02:30:37 +0000 (0:00:01.495)       0:04:54.073 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nThursday 26 March 2026  02:30:37 +0000 (0:00:00.480)       0:04:54.553 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nThursday 26 March 2026  02:30:38 +0000 (0:00:01.393)       0:04:55.946 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nThursday 26 March 2026  02:30:40 +0000 (0:00:01.863)       0:04:57.810 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nThursday 26 March 2026  02:30:43 +0000 (0:00:02.204)       0:05:00.014 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nThursday 26 March 2026  02:30:43 +0000 (0:00:00.433)       0:05:00.448 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nThursday 26 March 2026  02:30:44 +0000 (0:00:01.000)       0:05:01.448 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nThursday 26 March 2026  02:30:44 +0000 (0:00:00.502)       0:05:01.951 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nThursday 26 March 2026  02:30:46 +0000 (0:00:01.439)       0:05:03.390 ******** \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nThursday 26 March 2026  02:30:47 +0000 (0:00:01.611)       0:05:05.002 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nThursday 26 March 2026  02:30:54 +0000 (0:00:06.936)       0:05:11.938 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nThursday 26 March 2026  02:30:55 +0000 (0:00:00.770)       0:05:12.709 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nThursday 26 March 2026  02:30:56 +0000 (0:00:00.933)       0:05:13.642 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nThursday 26 March 2026  02:30:58 +0000 (0:00:01.400)       0:05:15.043 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : swift] ****************************************************\nThursday 26 March 2026  02:31:00 +0000 (0:00:02.290)       0:05:17.333 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nThursday 26 March 2026  02:31:00 +0000 (0:00:00.665)       0:05:17.999 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nThursday 26 March 2026  02:31:01 +0000 (0:00:00.356)       0:05:18.356 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nThursday 26 March 2026  02:31:01 +0000 (0:00:00.438)       0:05:18.794 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nThursday 26 March 2026  02:31:02 +0000 (0:00:00.337)       0:05:19.132 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nThursday 26 March 2026  02:31:02 +0000 (0:00:00.680)       0:05:19.812 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nThursday 26 March 2026  02:31:03 +0000 (0:00:00.602)       0:05:20.414 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nThursday 26 March 2026  02:31:04 +0000 (0:00:00.716)       0:05:21.130 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nThursday 26 March 2026  02:31:04 +0000 (0:00:00.729)       0:05:21.860 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nThursday 26 March 2026  02:31:05 +0000 (0:00:00.892)       0:05:22.753 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nThursday 26 March 2026  02:31:06 +0000 (0:00:00.969)       0:05:23.722 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nThursday 26 March 2026  02:31:07 +0000 (0:00:00.842)       0:05:24.565 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nThursday 26 March 2026  02:31:12 +0000 (0:00:05.397)       0:05:29.962 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nThursday 26 March 2026  02:31:16 +0000 (0:00:03.222)       0:05:33.185 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nThursday 26 March 2026  02:31:27 +0000 (0:00:10.974)       0:05:44.159 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nThursday 26 March 2026  02:31:31 +0000 (0:00:04.647)       0:05:48.807 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nThursday 26 March 2026  02:31:36 +0000 (0:00:04.277)       0:05:53.085 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nThursday 26 March 2026  02:31:36 +0000 (0:00:00.761)       0:05:53.847 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nThursday 26 March 2026  02:31:37 +0000 (0:00:00.403)       0:05:54.251 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nThursday 26 March 2026  02:31:37 +0000 (0:00:00.388)       0:05:54.639 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nThursday 26 March 2026  02:31:38 +0000 (0:00:00.398)       0:05:55.038 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nThursday 26 March 2026  02:31:38 +0000 (0:00:00.737)       0:05:55.775 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nThursday 26 March 2026  02:31:39 +0000 (0:00:00.373)       0:05:56.148 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nThursday 26 March 2026  02:31:43 +0000 (0:00:04.783)       0:06:00.932 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:31:44 +0000 (0:00:00.819)       0:06:01.751 ******** \n=============================================================================== \nloadbalancer : Start backup proxysql container ------------------------- 10.97s\nhaproxy-config : Copying over skyline haproxy config -------------------- 6.94s\nloadbalancer : Start backup haproxy container --------------------------- 5.40s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 5.35s\nloadbalancer : Wait for haproxy to listen on VIP ------------------------ 4.78s\nloadbalancer : Wait for backup proxysql to start ------------------------ 4.65s\nhaproxy-config : Copying over glance haproxy config --------------------- 4.52s\nhaproxy-config : Copying over neutron haproxy config -------------------- 4.36s\nhaproxy-config : Configuring firewall for glance ------------------------ 4.35s\nloadbalancer : Start backup keepalived container ------------------------ 4.28s\nhaproxy-config : Copying over nova haproxy config ----------------------- 4.20s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 4.19s\nhaproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.09s\nhaproxy-config : Copying over manila haproxy config --------------------- 3.96s\nhaproxy-config : Copying over designate haproxy config ------------------ 3.72s\nhaproxy-config : Copying over placement haproxy config ------------------ 3.61s\nhaproxy-config : Copying over magnum haproxy config --------------------- 3.55s\nhaproxy-config : Copying over cinder haproxy config --------------------- 3.49s\nhaproxy-config : Copying over barbican haproxy config ------------------- 3.40s\nservice-cert-copy : loadbalancer | Copying over extra CA certificates --- 3.35s\n\u001b[32m2026-03-26 02:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d9981432-5afc-4da2-8289-2d0601443640 (opensearch) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d9981432-5afc-4da2-8289-2d0601443640 (opensearch) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:31:52 +0000 (0:00:00.269)       0:00:00.269 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:31:52 +0000 (0:00:00.298)       0:00:00.568 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  02:31:53 +0000 (0:00:00.447)       0:00:01.015 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nThursday 26 March 2026  02:31:53 +0000 (0:00:00.538)       0:00:01.554 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nThursday 26 March 2026  02:31:54 +0000 (0:00:00.667)       0:00:02.221 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  02:31:56 +0000 (0:00:01.732)       0:00:03.953 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nThursday 26 March 2026  02:31:56 +0000 (0:00:00.571)       0:00:04.525 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:31:59 +0000 (0:00:02.402)       0:00:06.927 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nThursday 26 March 2026  02:32:00 +0000 (0:00:00.819)       0:00:07.747 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nThursday 26 March 2026  02:32:01 +0000 (0:00:01.091)       0:00:08.838 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nThursday 26 March 2026  02:32:03 +0000 (0:00:02.510)       0:00:11.349 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nThursday 26 March 2026  02:32:06 +0000 (0:00:02.379)       0:00:13.729 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Check opensearch containers] ********************************\nThursday 26 March 2026  02:32:07 +0000 (0:00:01.915)       0:00:15.644 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  02:32:09 +0000 (0:00:01.657)       0:00:17.302 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  02:32:09 +0000 (0:00:00.299)       0:00:17.601 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  02:32:10 +0000 (0:00:00.080)       0:00:17.682 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  02:32:10 +0000 (0:00:00.067)       0:00:17.750 ******** \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nThursday 26 March 2026  02:32:10 +0000 (0:00:00.090)       0:00:17.840 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nThursday 26 March 2026  02:32:10 +0000 (0:00:00.226)       0:00:18.066 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nThursday 26 March 2026  02:32:11 +0000 (0:00:00.683)       0:00:18.750 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nThursday 26 March 2026  02:33:22 +0000 (0:01:10.951)       0:01:29.701 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  02:34:46 +0000 (0:01:24.336)       0:02:54.038 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nThursday 26 March 2026  02:34:46 +0000 (0:00:00.522)       0:02:54.561 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nThursday 26 March 2026  02:34:49 +0000 (0:00:02.979)       0:02:57.540 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nThursday 26 March 2026  02:34:52 +0000 (0:00:02.245)       0:02:59.786 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nThursday 26 March 2026  02:34:54 +0000 (0:00:02.731)       0:03:02.518 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:34:57 +0000 (0:00:02.568)       0:03:05.086 ******** \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 84.34s\nopensearch : Restart opensearch container ------------------------------ 70.95s\nopensearch : Wait for OpenSearch to become ready ------------------------ 2.98s\nopensearch : Create new log retention policy ---------------------------- 2.73s\nopensearch : Apply retention policy to existing indices ----------------- 2.57s\nopensearch : Copying over config.json files for services ---------------- 2.51s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 2.40s\nopensearch : Copying over opensearch service config file ---------------- 2.38s\nopensearch : Check if a log retention policy exists --------------------- 2.25s\nopensearch : Copying over opensearch-dashboards config file ------------- 1.92s\nopensearch : Ensuring config directories exist -------------------------- 1.73s\nopensearch : Check opensearch containers -------------------------------- 1.66s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 1.09s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 0.82s\nopensearch : Perform a flush -------------------------------------------- 0.68s\nopensearch : Setting sysctl values -------------------------------------- 0.67s\nopensearch : include_tasks ---------------------------------------------- 0.57s\nopensearch : include_tasks ---------------------------------------------- 0.54s\nopensearch : include_tasks ---------------------------------------------- 0.52s\nGroup hosts based on enabled services ----------------------------------- 0.45s\n\u001b[32m2026-03-26 02:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e3d112e1-7037-4de8-b783-8f3076c4424e (memcached) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e3d112e1-7037-4de8-b783-8f3076c4424e (memcached) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:35:04 +0000 (0:00:00.281)       0:00:00.281 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:35:04 +0000 (0:00:00.362)       0:00:00.644 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nThursday 26 March 2026  02:35:05 +0000 (0:00:00.469)       0:00:01.113 ******** \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nThursday 26 March 2026  02:35:05 +0000 (0:00:00.498)       0:00:01.611 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nThursday 26 March 2026  02:35:06 +0000 (0:00:00.681)       0:00:02.293 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\nTASK [memcached : Check memcached container] ***********************************\nThursday 26 March 2026  02:35:08 +0000 (0:00:01.774)       0:00:04.067 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nThursday 26 March 2026  02:35:09 +0000 (0:00:01.577)       0:00:05.645 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:35:12 +0000 (0:00:02.122)       0:00:07.767 ******** \n=============================================================================== \nmemcached : Restart memcached container --------------------------------- 2.12s\nmemcached : Copying over config.json files for services ----------------- 1.77s\nmemcached : Check memcached container ----------------------------------- 1.58s\nmemcached : Ensuring config directories exist --------------------------- 0.68s\nmemcached : include_tasks ----------------------------------------------- 0.50s\nGroup hosts based on enabled services ----------------------------------- 0.47s\nGroup hosts based on Kolla action --------------------------------------- 0.36s\n\u001b[32m2026-03-26 02:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a877614-1e0e-4232-a1a6-b271c8628b0b (redis) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3a877614-1e0e-4232-a1a6-b271c8628b0b (redis) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:35:19 +0000 (0:00:00.275)       0:00:00.275 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:35:19 +0000 (0:00:00.333)       0:00:00.609 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nThursday 26 March 2026  02:35:20 +0000 (0:00:00.433)       0:00:01.042 ******** \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nThursday 26 March 2026  02:35:20 +0000 (0:00:00.491)       0:00:01.534 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nThursday 26 March 2026  02:35:21 +0000 (0:00:01.076)       0:00:02.610 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nThursday 26 March 2026  02:35:24 +0000 (0:00:02.523)       0:00:05.134 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Check redis containers] ******************************************\nThursday 26 March 2026  02:35:26 +0000 (0:00:02.558)       0:00:07.692 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  02:35:28 +0000 (0:00:01.467)       0:00:09.159 ******** \n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  02:35:28 +0000 (0:00:00.068)       0:00:09.227 ******** \n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  02:35:28 +0000 (0:00:00.069)       0:00:09.297 ******** \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nThursday 26 March 2026  02:35:28 +0000 (0:00:00.067)       0:00:09.364 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nThursday 26 March 2026  02:35:36 +0000 (0:00:07.569)       0:00:16.934 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:35:44 +0000 (0:00:08.438)       0:00:25.373 ******** \n=============================================================================== \nredis : Restart redis-sentinel container -------------------------------- 8.44s\nredis : Restart redis container ----------------------------------------- 7.57s\nredis : Copying over redis config files --------------------------------- 2.56s\nredis : Copying over default config.json files -------------------------- 2.52s\nredis : Check redis containers ------------------------------------------ 1.47s\nredis : Ensuring config directories exist ------------------------------- 1.08s\nredis : include_tasks --------------------------------------------------- 0.49s\nGroup hosts based on enabled services ----------------------------------- 0.43s\nGroup hosts based on Kolla action --------------------------------------- 0.33s\nredis : Flush handlers -------------------------------------------------- 0.21s\n\u001b[32m2026-03-26 02:35:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33359332-80a6-46d1-87b5-7308d2e3af07 (mariadb) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:35:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 33359332-80a6-46d1-87b5-7308d2e3af07 (mariadb) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:35:51 +0000 (0:00:00.167)       0:00:00.167 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:35:52 +0000 (0:00:00.346)       0:00:00.514 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nThursday 26 March 2026  02:35:52 +0000 (0:00:00.577)       0:00:01.091 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  02:35:53 +0000 (0:00:00.376)       0:00:01.468 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nThursday 26 March 2026  02:35:53 +0000 (0:00:00.568)       0:00:02.036 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nThursday 26 March 2026  02:35:56 +0000 (0:00:02.874)       0:00:04.911 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nThursday 26 March 2026  02:35:57 +0000 (0:00:00.647)       0:00:05.559 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nThursday 26 March 2026  02:35:58 +0000 (0:00:01.395)       0:00:06.954 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nThursday 26 March 2026  02:36:01 +0000 (0:00:03.262)       0:00:10.217 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nThursday 26 March 2026  02:36:02 +0000 (0:00:01.090)       0:00:11.307 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  02:36:06 +0000 (0:00:03.848)       0:00:15.156 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nThursday 26 March 2026  02:36:07 +0000 (0:00:00.552)       0:00:15.708 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  02:36:09 +0000 (0:00:02.272)       0:00:17.980 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nThursday 26 March 2026  02:36:12 +0000 (0:00:02.584)       0:00:20.565 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check mariadb containers] **************************************\nThursday 26 March 2026  02:36:14 +0000 (0:00:02.268)       0:00:22.833 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nThursday 26 March 2026  02:36:17 +0000 (0:00:02.683)       0:00:25.517 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nThursday 26 March 2026  02:36:17 +0000 (0:00:00.831)       0:00:26.348 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nThursday 26 March 2026  02:36:18 +0000 (0:00:00.560)       0:00:26.908 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nThursday 26 March 2026  02:36:18 +0000 (0:00:00.356)       0:00:27.265 ******** \n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nThursday 26 March 2026  02:36:29 +0000 (0:00:10.949)       0:00:38.215 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nThursday 26 March 2026  02:36:30 +0000 (0:00:00.426)       0:00:38.641 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nThursday 26 March 2026  02:36:30 +0000 (0:00:00.685)       0:00:39.326 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nThursday 26 March 2026  02:36:31 +0000 (0:00:00.441)       0:00:39.767 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nThursday 26 March 2026  02:36:31 +0000 (0:00:00.424)       0:00:40.192 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nThursday 26 March 2026  02:36:32 +0000 (0:00:00.431)       0:00:40.623 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  02:36:33 +0000 (0:00:00.915)       0:00:41.539 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [mariadb : Running MariaDB bootstrap container] ***************************\nThursday 26 March 2026  02:36:33 +0000 (0:00:00.400)       0:00:41.939 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Store bootstrap host name into facts] **************************\nThursday 26 March 2026  02:36:43 +0000 (0:00:10.184)       0:00:52.124 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  02:36:43 +0000 (0:00:00.133)       0:00:52.257 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Starting first MariaDB container] *******************\nThursday 26 March 2026  02:36:44 +0000 (0:00:01.086)       0:00:53.344 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******\nThursday 26 March 2026  02:36:52 +0000 (0:00:07.778)       0:01:01.123 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******\nThursday 26 March 2026  02:36:54 +0000 (0:00:01.580)       0:01:02.703 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***\nThursday 26 March 2026  02:36:56 +0000 (0:00:02.667)       0:01:05.370 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nThursday 26 March 2026  02:36:57 +0000 (0:00:00.144)       0:01:05.515 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nThursday 26 March 2026  02:36:57 +0000 (0:00:00.327)       0:01:05.843 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  02:36:57 +0000 (0:00:00.573)       0:01:06.416 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  02:37:16 +0000 (0:00:18.300)       0:01:24.717 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  02:37:32 +0000 (0:00:16.579)       0:01:41.296 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  02:37:35 +0000 (0:00:02.459)       0:01:43.756 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  02:37:53 +0000 (0:00:18.549)       0:02:02.306 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  02:38:10 +0000 (0:00:16.528)       0:02:18.834 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  02:38:12 +0000 (0:00:02.554)       0:02:21.389 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  02:38:25 +0000 (0:00:12.060)       0:02:33.449 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  02:38:29 +0000 (0:00:04.563)       0:02:38.012 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nThursday 26 March 2026  02:38:32 +0000 (0:00:02.795)       0:02:40.808 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nThursday 26 March 2026  02:38:32 +0000 (0:00:00.583)       0:02:41.392 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nThursday 26 March 2026  02:38:35 +0000 (0:00:02.113)       0:02:43.505 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nThursday 26 March 2026  02:38:36 +0000 (0:00:01.906)       0:02:45.411 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nThursday 26 March 2026  02:38:39 +0000 (0:00:02.185)       0:02:47.596 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nThursday 26 March 2026  02:38:41 +0000 (0:00:01.850)       0:02:49.447 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nThursday 26 March 2026  02:38:43 +0000 (0:00:02.833)       0:02:52.281 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:38:44 +0000 (0:00:00.435)       0:02:52.716 ******** \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 36.85s\nmariadb : Wait for MariaDB service port liveness ----------------------- 33.11s\nmariadb : Restart MariaDB container ------------------------------------ 12.06s\nmariadb : Check MariaDB service port liveness -------------------------- 10.95s\nmariadb : Running MariaDB bootstrap container -------------------------- 10.18s\nmariadb : Starting first MariaDB container ------------------------------ 7.78s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.01s\nmariadb : Wait for MariaDB service port liveness ------------------------ 4.56s\nmariadb : Copying over galera.cnf --------------------------------------- 3.85s\nmariadb : Copying over config.json files for services ------------------- 3.26s\nmariadb : Ensuring config directories exist ----------------------------- 2.87s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 2.83s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.80s\nmariadb : Check mariadb containers -------------------------------------- 2.68s\nmariadb : Wait for first MariaDB service to sync WSREP ------------------ 2.67s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.58s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 2.27s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.27s\nmariadb : Creating database backup user and setting permissions --------- 2.19s\nmariadb : Creating shard root mysql user -------------------------------- 2.11s\n\u001b[32m2026-03-26 02:38:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639b807-1d34-43fe-916c-01e997ccc45d (rabbitmq) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:38:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b639b807-1d34-43fe-916c-01e997ccc45d (rabbitmq) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:38:51 +0000 (0:00:00.193)       0:00:00.193 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:38:51 +0000 (0:00:00.330)       0:00:00.524 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  02:38:52 +0000 (0:00:00.572)       0:00:01.096 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nThursday 26 March 2026  02:38:52 +0000 (0:00:00.550)       0:00:01.646 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nThursday 26 March 2026  02:38:53 +0000 (0:00:00.918)       0:00:02.565 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nThursday 26 March 2026  02:38:54 +0000 (0:00:00.371)       0:00:02.936 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nThursday 26 March 2026  02:38:54 +0000 (0:00:00.433)       0:00:03.370 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nThursday 26 March 2026  02:38:55 +0000 (0:00:00.416)       0:00:03.787 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  02:38:55 +0000 (0:00:00.609)       0:00:04.396 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nThursday 26 March 2026  02:38:56 +0000 (0:00:00.959)       0:00:05.356 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nThursday 26 March 2026  02:38:57 +0000 (0:00:00.781)       0:00:06.137 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nThursday 26 March 2026  02:38:57 +0000 (0:00:00.365)       0:00:06.502 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nThursday 26 March 2026  02:38:58 +0000 (0:00:00.371)       0:00:06.874 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nThursday 26 March 2026  02:38:59 +0000 (0:00:00.826)       0:00:07.700 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nThursday 26 March 2026  02:39:00 +0000 (0:00:01.567)       0:00:09.267 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nThursday 26 March 2026  02:39:01 +0000 (0:00:01.319)       0:00:10.586 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nThursday 26 March 2026  02:39:03 +0000 (0:00:01.714)       0:00:12.301 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nThursday 26 March 2026  02:39:05 +0000 (0:00:01.410)       0:00:13.711 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nThursday 26 March 2026  02:39:06 +0000 (0:00:01.755)       0:00:15.467 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nThursday 26 March 2026  02:39:08 +0000 (0:00:01.430)       0:00:16.898 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  02:39:09 +0000 (0:00:01.454)       0:00:18.352 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Check rabbitmq containers] ************************************\nThursday 26 March 2026  02:39:10 +0000 (0:00:00.405)       0:00:18.757 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Creating rabbitmq volume] *************************************\nThursday 26 March 2026  02:39:11 +0000 (0:00:01.184)       0:00:19.942 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Running RabbitMQ bootstrap container] *************************\nThursday 26 March 2026  02:39:12 +0000 (0:00:00.884)       0:00:20.826 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nThursday 26 March 2026  02:39:19 +0000 (0:00:07.279)       0:00:28.106 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  02:39:20 +0000 (0:00:00.594)       0:00:28.701 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  02:39:20 +0000 (0:00:00.614)       0:00:29.315 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  02:39:20 +0000 (0:00:00.245)       0:00:29.560 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  02:39:27 +0000 (0:00:06.599)       0:00:36.160 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  02:40:14 +0000 (0:00:46.534)       0:01:22.695 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  02:40:14 +0000 (0:00:00.586)       0:01:23.282 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  02:40:14 +0000 (0:00:00.254)       0:01:23.536 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  02:40:16 +0000 (0:00:01.691)       0:01:25.228 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  02:40:30 +0000 (0:00:13.585)       0:01:38.813 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  02:40:30 +0000 (0:00:00.747)       0:01:39.561 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  02:40:31 +0000 (0:00:00.235)       0:01:39.796 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  02:40:32 +0000 (0:00:01.470)       0:01:41.266 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nThursday 26 March 2026  02:40:46 +0000 (0:00:14.155)       0:01:55.422 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nThursday 26 March 2026  02:40:47 +0000 (0:00:00.608)       0:01:56.030 ******** \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35menable_outward_rabbitmq_True\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35moutward_rabbitmq_restart\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role rabbitmq (outward)] *******************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Restart rabbitmq (outward) services] *************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply rabbitmq (outward) post-configuration] *****************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:40:49 +0000 (0:00:02.403)       0:01:58.434 ******** \n=============================================================================== \nrabbitmq : Waiting for rabbitmq to start ------------------------------- 74.28s\nrabbitmq : Restart rabbitmq container ----------------------------------- 9.76s\nrabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.28s\nrabbitmq : Enable all stable feature flags ------------------------------ 2.40s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 1.95s\nrabbitmq : Copying over advanced.config --------------------------------- 1.76s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 1.71s\nrabbitmq : Copying over config.json files for services ------------------ 1.57s\nrabbitmq : Copying over enabled_plugins --------------------------------- 1.45s\nrabbitmq : Copying over definitions.json -------------------------------- 1.43s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 1.41s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.32s\nrabbitmq : Check rabbitmq containers ------------------------------------ 1.18s\nrabbitmq : include_tasks ------------------------------------------------ 0.96s\nrabbitmq : Get container facts ------------------------------------------ 0.92s\nrabbitmq : Creating rabbitmq volume ------------------------------------- 0.88s\nrabbitmq : Ensuring config directories exist ---------------------------- 0.83s\nrabbitmq : Get container facts ------------------------------------------ 0.78s\nrabbitmq : Put RabbitMQ node into maintenance mode ---------------------- 0.74s\nrabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 0.61s\n\u001b[32m2026-03-26 02:40:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9d84b916-885e-4b98-a059-63c4a30b9927 (openvswitch) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:40:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9d84b916-885e-4b98-a059-63c4a30b9927 (openvswitch) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:40:57 +0000 (0:00:00.268)       0:00:00.268 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:40:57 +0000 (0:00:00.714)       0:00:00.983 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nThursday 26 March 2026  02:40:58 +0000 (0:00:00.638)       0:00:01.621 ******** \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nThursday 26 March 2026  02:40:59 +0000 (0:00:01.168)       0:00:02.790 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nThursday 26 March 2026  02:41:00 +0000 (0:00:01.219)       0:00:04.010 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nThursday 26 March 2026  02:41:02 +0000 (0:00:01.466)       0:00:05.477 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nThursday 26 March 2026  02:41:03 +0000 (0:00:01.217)       0:00:06.694 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nThursday 26 March 2026  02:41:04 +0000 (0:00:00.761)       0:00:07.456 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nThursday 26 March 2026  02:41:05 +0000 (0:00:01.473)       0:00:08.929 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nThursday 26 March 2026  02:41:08 +0000 (0:00:02.203)       0:00:11.133 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Check openvswitch containers] ******************************\nThursday 26 March 2026  02:41:09 +0000 (0:00:00.982)       0:00:12.115 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:10 +0000 (0:00:01.614)       0:00:13.729 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.330)       0:00:14.060 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.136)       0:00:14.196 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.144)       0:00:14.340 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.153)       0:00:14.494 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.148)       0:00:14.642 ******** \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nThursday 26 March 2026  02:41:11 +0000 (0:00:00.133)       0:00:14.776 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nThursday 26 March 2026  02:41:20 +0000 (0:00:08.407)       0:00:23.183 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nThursday 26 March 2026  02:41:21 +0000 (0:00:01.784)       0:00:24.968 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nThursday 26 March 2026  02:41:29 +0000 (0:00:08.005)       0:00:32.973 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nThursday 26 March 2026  02:41:36 +0000 (0:00:06.330)       0:00:39.304 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nThursday 26 March 2026  02:41:38 +0000 (0:00:02.424)       0:00:41.728 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nThursday 26 March 2026  02:41:41 +0000 (0:00:03.021)       0:00:44.750 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:41:48 +0000 (0:00:07.104)       0:00:51.855 ******** \n=============================================================================== \nopenvswitch : Restart openvswitch-vswitchd container ------------------- 15.11s\nopenvswitch : Restart openvswitch-db-server container ------------------- 8.41s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 6.33s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 3.02s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 2.42s\nopenvswitch : Copying over config.json files for services --------------- 2.20s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 1.78s\nopenvswitch : Check openvswitch containers ------------------------------ 1.61s\nopenvswitch : Ensuring config directories exist ------------------------- 1.47s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.47s\nmodule-load : Load modules ---------------------------------------------- 1.22s\nmodule-load : Drop module persistence ----------------------------------- 1.22s\nopenvswitch : include_tasks --------------------------------------------- 1.17s\nopenvswitch : Flush Handlers -------------------------------------------- 1.05s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.98s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 0.76s\nGroup hosts based on Kolla action --------------------------------------- 0.72s\nGroup hosts based on enabled services ----------------------------------- 0.64s\n\u001b[32m2026-03-26 02:41:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 13a853e3-ac1d-41a3-984d-fc24ff8bacb1 (ovn) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:41:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 13a853e3-ac1d-41a3-984d-fc24ff8bacb1 (ovn) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  02:41:56 +0000 (0:00:00.203)       0:00:00.203 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  02:41:57 +0000 (0:00:00.774)       0:00:00.978 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nThursday 26 March 2026  02:41:57 +0000 (0:00:00.867)       0:00:01.845 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nThursday 26 March 2026  02:41:59 +0000 (0:00:01.142)       0:00:02.988 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nThursday 26 March 2026  02:42:00 +0000 (0:00:01.188)       0:00:04.177 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nThursday 26 March 2026  02:42:01 +0000 (0:00:01.471)       0:00:05.648 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nThursday 26 March 2026  02:42:02 +0000 (0:00:01.271)       0:00:06.920 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Check ovn-controller containers] ************************\nThursday 26 March 2026  02:42:04 +0000 (0:00:01.634)       0:00:08.554 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nThursday 26 March 2026  02:42:05 +0000 (0:00:01.357)       0:00:09.912 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nThursday 26 March 2026  02:42:08 +0000 (0:00:02.333)       0:00:12.246 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:26 +0000 (0:00:18.232)       0:00:30.479 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:26 +0000 (0:00:00.260)       0:00:30.739 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:26 +0000 (0:00:00.105)       0:00:30.844 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:26 +0000 (0:00:00.098)       0:00:30.943 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:27 +0000 (0:00:00.071)       0:00:31.014 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  02:42:27 +0000 (0:00:00.073)       0:00:31.087 ******** \n\nRUNNING HANDLER [ovn-controller : Reload systemd config] ***********************\nThursday 26 March 2026  02:42:27 +0000 (0:00:00.073)       0:00:31.160 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nThursday 26 March 2026  02:42:28 +0000 (0:00:01.581)       0:00:32.741 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  02:43:01 +0000 (0:00:32.834)       0:01:05.575 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  02:43:02 +0000 (0:00:00.730)       0:01:06.306 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nThursday 26 March 2026  02:43:02 +0000 (0:00:00.534)       0:01:06.840 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nThursday 26 March 2026  02:43:03 +0000 (0:00:01.034)       0:01:07.875 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nThursday 26 March 2026  02:43:04 +0000 (0:00:00.338)       0:01:08.214 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nThursday 26 March 2026  02:43:04 +0000 (0:00:00.339)       0:01:08.553 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nThursday 26 March 2026  02:43:04 +0000 (0:00:00.333)       0:01:08.887 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nThursday 26 March 2026  02:43:05 +0000 (0:00:00.568)       0:01:09.456 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nThursday 26 March 2026  02:43:05 +0000 (0:00:00.322)       0:01:09.778 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nThursday 26 March 2026  02:43:06 +0000 (0:00:00.343)       0:01:10.122 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nThursday 26 March 2026  02:43:06 +0000 (0:00:00.338)       0:01:10.461 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nThursday 26 March 2026  02:43:06 +0000 (0:00:00.308)       0:01:10.769 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nThursday 26 March 2026  02:43:07 +0000 (0:00:00.547)       0:01:11.316 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nThursday 26 March 2026  02:43:07 +0000 (0:00:00.303)       0:01:11.620 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nThursday 26 March 2026  02:43:07 +0000 (0:00:00.309)       0:01:11.930 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nThursday 26 March 2026  02:43:08 +0000 (0:00:00.292)       0:01:12.222 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nThursday 26 March 2026  02:43:08 +0000 (0:00:00.506)       0:01:12.729 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nThursday 26 March 2026  02:43:09 +0000 (0:00:00.309)       0:01:13.039 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nThursday 26 March 2026  02:43:09 +0000 (0:00:00.303)       0:01:13.342 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  02:43:09 +0000 (0:00:00.305)       0:01:13.648 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************\nThursday 26 March 2026  02:43:10 +0000 (0:00:00.763)       0:01:14.411 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************\nThursday 26 March 2026  02:43:10 +0000 (0:00:00.442)       0:01:14.854 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check NB cluster status] ****************************************\nThursday 26 March 2026  02:43:11 +0000 (0:00:00.423)       0:01:15.278 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check SB cluster status] ****************************************\nThursday 26 March 2026  02:43:11 +0000 (0:00:00.334)       0:01:15.613 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***\nThursday 26 March 2026  02:43:12 +0000 (0:00:00.602)       0:01:16.216 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***\nThursday 26 March 2026  02:43:12 +0000 (0:00:00.372)       0:01:16.588 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************\nThursday 26 March 2026  02:43:12 +0000 (0:00:00.349)       0:01:16.938 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************\nThursday 26 March 2026  02:43:13 +0000 (0:00:00.330)       0:01:17.269 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nThursday 26 March 2026  02:43:13 +0000 (0:00:00.569)       0:01:17.839 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nThursday 26 March 2026  02:43:15 +0000 (0:00:01.476)       0:01:19.315 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nThursday 26 March 2026  02:43:19 +0000 (0:00:03.899)       0:01:23.215 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:21 +0000 (0:00:02.092)       0:01:25.307 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:21 +0000 (0:00:00.065)       0:01:25.373 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:21 +0000 (0:00:00.289)       0:01:25.662 ******** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nThursday 26 March 2026  02:43:21 +0000 (0:00:00.072)       0:01:25.735 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nThursday 26 March 2026  02:43:29 +0000 (0:00:07.747)       0:01:33.482 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nThursday 26 March 2026  02:43:32 +0000 (0:00:02.620)       0:01:36.103 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nThursday 26 March 2026  02:43:39 +0000 (0:00:07.646)       0:01:43.749 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nThursday 26 March 2026  02:43:39 +0000 (0:00:00.119)       0:01:43.869 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nThursday 26 March 2026  02:43:40 +0000 (0:00:01.058)       0:01:44.928 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nThursday 26 March 2026  02:43:41 +0000 (0:00:00.619)       0:01:45.547 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nThursday 26 March 2026  02:43:42 +0000 (0:00:00.798)       0:01:46.346 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nThursday 26 March 2026  02:43:43 +0000 (0:00:00.638)       0:01:46.984 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nThursday 26 March 2026  02:43:44 +0000 (0:00:01.219)       0:01:48.204 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Unset bootstrap args fact] **************************************\nThursday 26 March 2026  02:43:45 +0000 (0:00:00.766)       0:01:48.971 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nThursday 26 March 2026  02:43:45 +0000 (0:00:00.328)       0:01:49.299 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nThursday 26 March 2026  02:43:46 +0000 (0:00:01.476)       0:01:50.776 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nThursday 26 March 2026  02:43:50 +0000 (0:00:04.001)       0:01:54.777 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:54 +0000 (0:00:03.346)       0:01:58.124 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:54 +0000 (0:00:00.122)       0:01:58.246 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  02:43:54 +0000 (0:00:00.090)       0:01:58.337 ******** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nThursday 26 March 2026  02:43:54 +0000 (0:00:00.099)       0:01:58.436 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nThursday 26 March 2026  02:44:00 +0000 (0:00:06.304)       0:02:04.740 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nThursday 26 March 2026  02:44:07 +0000 (0:00:06.237)       0:02:10.978 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nThursday 26 March 2026  02:44:13 +0000 (0:00:06.207)       0:02:17.186 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nThursday 26 March 2026  02:44:13 +0000 (0:00:00.171)       0:02:17.357 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nThursday 26 March 2026  02:44:14 +0000 (0:00:01.063)       0:02:18.421 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nThursday 26 March 2026  02:44:15 +0000 (0:00:00.674)       0:02:19.096 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nThursday 26 March 2026  02:44:15 +0000 (0:00:00.822)       0:02:19.919 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nThursday 26 March 2026  02:44:16 +0000 (0:00:00.641)       0:02:20.561 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nThursday 26 March 2026  02:44:17 +0000 (0:00:01.176)       0:02:21.737 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:44:18 +0000 (0:00:01.000)       0:02:22.738 ******** \n=============================================================================== \novn-controller : Restart ovn-controller container ---------------------- 32.83s\novn-controller : Configure OVN in OVSDB -------------------------------- 18.23s\novn-db : Restart ovn-nb-db container ----------------------------------- 14.05s\novn-db : Restart ovn-northd container ---------------------------------- 13.85s\novn-db : Restart ovn-sb-db container ------------------------------------ 8.86s\novn-db : Copying over config.json files for services -------------------- 4.00s\novn-db : Copying over config.json files for services -------------------- 3.90s\novn-db : Check ovn containers ------------------------------------------- 3.35s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.33s\novn-db : Check ovn containers ------------------------------------------- 2.09s\novn-controller : Copying over systemd override -------------------------- 1.63s\novn-controller : Reload systemd config ---------------------------------- 1.58s\novn-db : Ensuring config directories exist ------------------------------ 1.48s\novn-db : Ensuring config directories exist ------------------------------ 1.48s\novn-controller : Copying over config.json files for services ------------ 1.47s\novn-controller : Check ovn-controller containers ------------------------ 1.36s\novn-controller : Ensuring systemd override directory exists ------------- 1.27s\novn-db : Wait for ovn-nb-db --------------------------------------------- 1.22s\novn-controller : Ensuring config directories exist ---------------------- 1.19s\novn-db : Wait for ovn-nb-db --------------------------------------------- 1.18s\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh\n\u001b[32m2026-03-26 02:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m\n\u001b[32m2026-03-26 02:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 89ea481d-9381-42c6-ad2c-7d8c23c0aae3 (wipe-partitions) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 89ea481d-9381-42c6-ad2c-7d8c23c0aae3 (wipe-partitions) has been started and output is visible here.\u001b[0m\n\nPLAY [Wipe partitions] *********************************************************\n\nTASK [Find all logical devices owned by UID 167] *******************************\nThursday 26 March 2026  02:44:36 +0000 (0:00:00.153)       0:00:00.153 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Remove all rook related logical devices] *********************************\nThursday 26 March 2026  02:44:36 +0000 (0:00:00.590)       0:00:00.744 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Find all logical devices with prefix ceph] *******************************\nThursday 26 March 2026  02:44:37 +0000 (0:00:00.398)       0:00:01.143 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Remove all ceph related logical devices] *********************************\nThursday 26 March 2026  02:44:38 +0000 (0:00:00.637)       0:00:01.780 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check device availability] ***********************************************\nThursday 26 March 2026  02:44:38 +0000 (0:00:00.281)       0:00:02.062 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Wipe partitions with wipefs] *********************************************\nThursday 26 March 2026  02:44:39 +0000 (0:00:01.215)       0:00:03.278 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Overwrite first 32M with zeros] ******************************************\nThursday 26 March 2026  02:44:41 +0000 (0:00:01.629)       0:00:04.907 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\nTASK [Reload udev rules] *******************************************************\nThursday 26 March 2026  02:44:44 +0000 (0:00:03.037)       0:00:07.944 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Request device events from the kernel] ***********************************\nThursday 26 March 2026  02:44:44 +0000 (0:00:00.649)       0:00:08.594 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:44:45 +0000 (0:00:00.645)       0:00:09.240 ******** \n=============================================================================== \nOverwrite first 32M with zeros ------------------------------------------ 3.04s\nWipe partitions with wipefs --------------------------------------------- 1.63s\nCheck device availability ----------------------------------------------- 1.22s\nReload udev rules ------------------------------------------------------- 0.65s\nRequest device events from the kernel ----------------------------------- 0.65s\nFind all logical devices with prefix ceph ------------------------------- 0.64s\nFind all logical devices owned by UID 167 ------------------------------- 0.59s\nRemove all rook related logical devices --------------------------------- 0.40s\nRemove all ceph related logical devices --------------------------------- 0.28s\n\u001b[32m2026-03-26 02:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a5e1532e-72fd-419b-9693-6cea10c48d87 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a5e1532e-72fd-419b-9693-6cea10c48d87 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nThursday 26 March 2026  02:45:03 +0000 (0:00:00.297)       0:00:00.297 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nThursday 26 March 2026  02:45:04 +0000 (0:00:01.137)       0:00:01.434 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:45:05 +0000 (0:00:01.370)       0:00:02.805 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nThursday 26 March 2026  02:45:11 +0000 (0:00:06.124)       0:00:08.929 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:45:12 +0000 (0:00:00.597)       0:00:09.527 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 6.12s\nosism.commons.facts : Copy fact files ----------------------------------- 1.37s\nosism.commons.facts : Create custom facts directory --------------------- 1.14s\nGather facts for all hosts ---------------------------------------------- 0.60s\n\u001b[32m2026-03-26 02:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b695aa9f-dffc-4083-94dc-70d9d0de54bc (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b695aa9f-dffc-4083-94dc-70d9d0de54bc (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:45:20 +0000 (0:00:00.371)       0:00:00.371 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:45:20 +0000 (0:00:00.281)       0:00:00.653 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:21 +0000 (0:00:00.270)       0:00:00.924 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:21 +0000 (0:00:00.553)       0:00:01.477 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:21 +0000 (0:00:00.248)       0:00:01.726 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:22 +0000 (0:00:00.224)       0:00:01.950 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:22 +0000 (0:00:00.206)       0:00:02.156 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:22 +0000 (0:00:00.215)       0:00:02.372 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:22 +0000 (0:00:00.218)       0:00:02.591 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:23 +0000 (0:00:00.245)       0:00:02.836 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:23 +0000 (0:00:00.220)       0:00:03.057 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:23 +0000 (0:00:00.226)       0:00:03.283 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:24 +0000 (0:00:00.708)       0:00:03.992 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:24 +0000 (0:00:00.720)       0:00:04.712 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:25 +0000 (0:00:01.090)       0:00:05.802 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:26 +0000 (0:00:00.496)       0:00:06.299 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:26 +0000 (0:00:00.382)       0:00:06.682 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:27 +0000 (0:00:00.446)       0:00:07.128 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:27 +0000 (0:00:00.238)       0:00:07.367 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:27 +0000 (0:00:00.216)       0:00:07.583 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:28 +0000 (0:00:00.228)       0:00:07.812 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:28 +0000 (0:00:00.222)       0:00:08.035 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:28 +0000 (0:00:00.244)       0:00:08.279 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:29 +0000 (0:00:00.681)       0:00:08.960 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:29 +0000 (0:00:00.256)       0:00:09.216 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:29 +0000 (0:00:00.211)       0:00:09.428 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:30 +0000 (0:00:00.783)       0:00:10.211 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:30 +0000 (0:00:00.233)       0:00:10.445 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:30 +0000 (0:00:00.232)       0:00:10.677 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:31 +0000 (0:00:00.216)       0:00:10.894 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nThursday 26 March 2026  02:45:31 +0000 (0:00:00.232)       0:00:11.126 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nThursday 26 March 2026  02:45:31 +0000 (0:00:00.200)       0:00:11.326 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nThursday 26 March 2026  02:45:31 +0000 (0:00:00.177)       0:00:11.504 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nThursday 26 March 2026  02:45:31 +0000 (0:00:00.203)       0:00:11.708 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nThursday 26 March 2026  02:45:32 +0000 (0:00:00.396)       0:00:12.105 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nThursday 26 March 2026  02:45:32 +0000 (0:00:00.159)       0:00:12.264 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nThursday 26 March 2026  02:45:32 +0000 (0:00:00.208)       0:00:12.473 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nThursday 26 March 2026  02:45:32 +0000 (0:00:00.161)       0:00:12.634 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.184)       0:00:12.818 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.175)       0:00:12.993 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.149)       0:00:13.142 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.144)       0:00:13.286 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.151)       0:00:13.438 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.139)       0:00:13.577 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nThursday 26 March 2026  02:45:33 +0000 (0:00:00.129)       0:00:13.707 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nThursday 26 March 2026  02:45:34 +0000 (0:00:00.391)       0:00:14.098 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nThursday 26 March 2026  02:45:34 +0000 (0:00:00.154)       0:00:14.254 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nThursday 26 March 2026  02:45:34 +0000 (0:00:00.134)       0:00:14.388 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nThursday 26 March 2026  02:45:34 +0000 (0:00:00.151)       0:00:14.540 ******** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-e2623153-bc41-510f-8884-ef957bb96082\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nThursday 26 March 2026  02:45:34 +0000 (0:00:00.232)       0:00:14.772 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:45:36 +0000 (0:00:01.998)       0:00:16.770 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:45:37 +0000 (0:00:00.269)       0:00:17.040 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:37 +0000 (0:00:00.241)       0:00:17.282 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:38 +0000 (0:00:00.652)       0:00:17.934 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:38 +0000 (0:00:00.239)       0:00:18.174 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:38 +0000 (0:00:00.243)       0:00:18.417 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:38 +0000 (0:00:00.228)       0:00:18.646 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:39 +0000 (0:00:00.221)       0:00:18.868 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:39 +0000 (0:00:00.233)       0:00:19.101 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:39 +0000 (0:00:00.210)       0:00:19.312 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:39 +0000 (0:00:00.212)       0:00:19.525 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:39 +0000 (0:00:00.212)       0:00:19.738 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:40 +0000 (0:00:00.700)       0:00:20.438 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:41 +0000 (0:00:00.735)       0:00:21.173 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:42 +0000 (0:00:00.966)       0:00:22.140 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:42 +0000 (0:00:00.480)       0:00:22.620 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:43 +0000 (0:00:00.368)       0:00:22.989 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:43 +0000 (0:00:00.424)       0:00:23.414 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:43 +0000 (0:00:00.219)       0:00:23.634 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:44 +0000 (0:00:00.208)       0:00:23.842 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:44 +0000 (0:00:00.206)       0:00:24.048 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:44 +0000 (0:00:00.230)       0:00:24.279 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:44 +0000 (0:00:00.220)       0:00:24.500 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:44 +0000 (0:00:00.218)       0:00:24.719 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:45 +0000 (0:00:00.212)       0:00:24.931 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:45 +0000 (0:00:00.726)       0:00:25.658 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:46 +0000 (0:00:00.727)       0:00:26.386 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:46 +0000 (0:00:00.221)       0:00:26.607 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.224)       0:00:26.832 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.270)       0:00:27.103 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.211)       0:00:27.314 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.179)       0:00:27.493 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.135)       0:00:27.629 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nThursday 26 March 2026  02:45:47 +0000 (0:00:00.141)       0:00:27.771 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nThursday 26 March 2026  02:45:48 +0000 (0:00:00.145)       0:00:27.916 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nThursday 26 March 2026  02:45:48 +0000 (0:00:00.130)       0:00:28.046 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nThursday 26 March 2026  02:45:48 +0000 (0:00:00.199)       0:00:28.246 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nThursday 26 March 2026  02:45:48 +0000 (0:00:00.397)       0:00:28.643 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nThursday 26 March 2026  02:45:48 +0000 (0:00:00.157)       0:00:28.800 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.152)       0:00:28.953 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.138)       0:00:29.091 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.149)       0:00:29.241 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.143)       0:00:29.385 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.141)       0:00:29.526 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nThursday 26 March 2026  02:45:49 +0000 (0:00:00.153)       0:00:29.680 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nThursday 26 March 2026  02:45:50 +0000 (0:00:00.147)       0:00:29.827 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nThursday 26 March 2026  02:45:50 +0000 (0:00:00.132)       0:00:29.960 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nThursday 26 March 2026  02:45:50 +0000 (0:00:00.140)       0:00:30.100 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nThursday 26 March 2026  02:45:50 +0000 (0:00:00.163)       0:00:30.264 ******** \n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-a652979e-9f40-503a-bbc8-6de5e605991e\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nThursday 26 March 2026  02:45:50 +0000 (0:00:00.467)       0:00:30.731 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:45:52 +0000 (0:00:01.224)       0:00:31.956 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:45:52 +0000 (0:00:00.289)       0:00:32.246 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:52 +0000 (0:00:00.275)       0:00:32.521 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:53 +0000 (0:00:00.401)       0:00:32.923 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:53 +0000 (0:00:00.203)       0:00:33.126 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:53 +0000 (0:00:00.198)       0:00:33.325 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:53 +0000 (0:00:00.234)       0:00:33.559 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:54 +0000 (0:00:00.703)       0:00:34.263 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:54 +0000 (0:00:00.228)       0:00:34.491 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:54 +0000 (0:00:00.219)       0:00:34.710 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:55 +0000 (0:00:00.226)       0:00:34.937 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:55 +0000 (0:00:00.205)       0:00:35.142 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:55 +0000 (0:00:00.444)       0:00:35.587 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:56 +0000 (0:00:00.502)       0:00:36.089 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:56 +0000 (0:00:00.453)       0:00:36.543 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:45:57 +0000 (0:00:00.511)       0:00:37.055 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:57 +0000 (0:00:00.407)       0:00:37.462 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:58 +0000 (0:00:00.682)       0:00:38.145 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:58 +0000 (0:00:00.204)       0:00:38.349 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:58 +0000 (0:00:00.228)       0:00:38.578 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:58 +0000 (0:00:00.211)       0:00:38.789 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:59 +0000 (0:00:00.253)       0:00:39.043 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:59 +0000 (0:00:00.207)       0:00:39.251 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:59 +0000 (0:00:00.227)       0:00:39.478 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:45:59 +0000 (0:00:00.219)       0:00:39.698 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:46:00 +0000 (0:00:00.236)       0:00:39.934 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:46:01 +0000 (0:00:00.952)       0:00:40.887 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:46:01 +0000 (0:00:00.188)       0:00:41.075 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:46:01 +0000 (0:00:00.194)       0:00:41.269 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:46:02 +0000 (0:00:00.762)       0:00:42.032 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nThursday 26 March 2026  02:46:02 +0000 (0:00:00.220)       0:00:42.252 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nThursday 26 March 2026  02:46:02 +0000 (0:00:00.197)       0:00:42.450 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nThursday 26 March 2026  02:46:02 +0000 (0:00:00.141)       0:00:42.591 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nThursday 26 March 2026  02:46:02 +0000 (0:00:00.148)       0:00:42.740 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.133)       0:00:42.874 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.168)       0:00:43.043 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.188)       0:00:43.231 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.152)       0:00:43.383 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.179)       0:00:43.562 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nThursday 26 March 2026  02:46:03 +0000 (0:00:00.155)       0:00:43.717 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nThursday 26 March 2026  02:46:04 +0000 (0:00:00.170)       0:00:43.888 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nThursday 26 March 2026  02:46:04 +0000 (0:00:00.393)       0:00:44.282 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nThursday 26 March 2026  02:46:04 +0000 (0:00:00.153)       0:00:44.435 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nThursday 26 March 2026  02:46:04 +0000 (0:00:00.152)       0:00:44.588 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nThursday 26 March 2026  02:46:04 +0000 (0:00:00.134)       0:00:44.723 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nThursday 26 March 2026  02:46:05 +0000 (0:00:00.156)       0:00:44.879 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nThursday 26 March 2026  02:46:05 +0000 (0:00:00.145)       0:00:45.025 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nThursday 26 March 2026  02:46:05 +0000 (0:00:00.145)       0:00:45.171 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nThursday 26 March 2026  02:46:05 +0000 (0:00:00.139)       0:00:45.310 ******** \n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nThursday 26 March 2026  02:46:05 +0000 (0:00:00.222)       0:00:45.533 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:46:06 +0000 (0:00:01.078)       0:00:46.612 ******** \n=============================================================================== \nWrite configuration file ------------------------------------------------ 4.30s\nAdd known links to the list of available block devices ------------------ 1.61s\nAdd known partitions to the list of available block devices ------------- 1.55s\nAdd known links to the list of available block devices ------------------ 1.09s\nAdd known links to the list of available block devices ------------------ 0.97s\nAdd known partitions to the list of available block devices ------------- 0.95s\nPrint configuration data ------------------------------------------------ 0.92s\nGet extra vars for Ceph configuration ----------------------------------- 0.84s\nGet initial list of available block devices ----------------------------- 0.79s\nAdd known partitions to the list of available block devices ------------- 0.78s\nAdd known partitions to the list of available block devices ------------- 0.76s\nAdd known links to the list of available block devices ------------------ 0.74s\nAdd known partitions to the list of available block devices ------------- 0.73s\nAdd known partitions to the list of available block devices ------------- 0.73s\nAdd known links to the list of available block devices ------------------ 0.72s\nGenerate lvm_volumes structure (block + db) ----------------------------- 0.71s\nAdd known links to the list of available block devices ------------------ 0.71s\nAdd known links to the list of available block devices ------------------ 0.70s\nAdd known links to the list of available block devices ------------------ 0.70s\nPrint ceph_osd_devices -------------------------------------------------- 0.70s\n\u001b[32m2026-03-26 02:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bfb64ee4-2646-4665-9bf0-88d97a04f762 (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-03-26 02:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-03-26 02:46:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-03-26 02:46:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-03-26 02:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-03-26 02:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n[master 7661ed7] 2026-03-26-02-47\n 1 file changed, 30 insertions(+), 9 deletions(-)\n\u001b[32m2026-03-26 02:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 84ec70a8-97fe-4cd7-9d64-7d2214c4453b (ceph-create-lvm-devices) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 84ec70a8-97fe-4cd7-9d64-7d2214c4453b (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:47:08 +0000 (0:00:00.339)       0:00:00.339 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:47:08 +0000 (0:00:00.268)       0:00:00.608 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:09 +0000 (0:00:00.308)       0:00:00.916 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:09 +0000 (0:00:00.593)       0:00:01.510 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:10 +0000 (0:00:00.251)       0:00:01.762 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:10 +0000 (0:00:00.234)       0:00:01.997 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:10 +0000 (0:00:00.243)       0:00:02.240 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:10 +0000 (0:00:00.245)       0:00:02.486 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:10 +0000 (0:00:00.225)       0:00:02.711 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:11 +0000 (0:00:00.236)       0:00:02.948 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:11 +0000 (0:00:00.229)       0:00:03.177 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:11 +0000 (0:00:00.225)       0:00:03.403 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:12 +0000 (0:00:00.713)       0:00:04.116 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:13 +0000 (0:00:00.717)       0:00:04.834 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:14 +0000 (0:00:00.979)       0:00:05.813 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:14 +0000 (0:00:00.477)       0:00:06.290 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:14 +0000 (0:00:00.334)       0:00:06.625 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:15 +0000 (0:00:00.461)       0:00:07.087 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:15 +0000 (0:00:00.231)       0:00:07.318 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:15 +0000 (0:00:00.218)       0:00:07.537 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:16 +0000 (0:00:00.212)       0:00:07.750 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:16 +0000 (0:00:00.214)       0:00:07.964 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:16 +0000 (0:00:00.203)       0:00:08.168 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:17 +0000 (0:00:00.670)       0:00:08.839 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:17 +0000 (0:00:00.209)       0:00:09.048 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:17 +0000 (0:00:00.211)       0:00:09.259 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:18 +0000 (0:00:00.731)       0:00:09.991 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:18 +0000 (0:00:00.247)       0:00:10.238 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:18 +0000 (0:00:00.241)       0:00:10.480 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:18 +0000 (0:00:00.229)       0:00:10.709 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nThursday 26 March 2026  02:47:19 +0000 (0:00:00.238)       0:00:10.948 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nThursday 26 March 2026  02:47:19 +0000 (0:00:00.154)       0:00:11.102 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nThursday 26 March 2026  02:47:19 +0000 (0:00:00.217)       0:00:11.320 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nThursday 26 March 2026  02:47:21 +0000 (0:00:02.059)       0:00:13.379 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nThursday 26 March 2026  02:47:22 +0000 (0:00:00.384)       0:00:13.764 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nThursday 26 March 2026  02:47:23 +0000 (0:00:01.510)       0:00:15.275 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nThursday 26 March 2026  02:47:23 +0000 (0:00:00.154)       0:00:15.430 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nThursday 26 March 2026  02:47:23 +0000 (0:00:00.157)       0:00:15.587 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.175)       0:00:15.763 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.146)       0:00:15.909 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.161)       0:00:16.071 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.141)       0:00:16.212 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.181)       0:00:16.393 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.156)       0:00:16.550 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nThursday 26 March 2026  02:47:24 +0000 (0:00:00.168)       0:00:16.719 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nThursday 26 March 2026  02:47:25 +0000 (0:00:00.406)       0:00:17.125 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nThursday 26 March 2026  02:47:25 +0000 (0:00:00.238)       0:00:17.364 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nThursday 26 March 2026  02:47:25 +0000 (0:00:00.143)       0:00:17.507 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nThursday 26 March 2026  02:47:25 +0000 (0:00:00.150)       0:00:17.657 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nThursday 26 March 2026  02:47:26 +0000 (0:00:00.151)       0:00:17.808 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nThursday 26 March 2026  02:47:26 +0000 (0:00:00.156)       0:00:17.965 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nThursday 26 March 2026  02:47:26 +0000 (0:00:00.154)       0:00:18.119 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nThursday 26 March 2026  02:47:26 +0000 (0:00:00.178)       0:00:18.298 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nThursday 26 March 2026  02:47:27 +0000 (0:00:00.701)       0:00:18.999 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nThursday 26 March 2026  02:47:27 +0000 (0:00:00.503)       0:00:19.502 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nThursday 26 March 2026  02:47:28 +0000 (0:00:00.523)       0:00:20.026 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nThursday 26 March 2026  02:47:28 +0000 (0:00:00.389)       0:00:20.416 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nThursday 26 March 2026  02:47:28 +0000 (0:00:00.120)       0:00:20.536 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nThursday 26 March 2026  02:47:28 +0000 (0:00:00.126)       0:00:20.663 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.147)       0:00:20.811 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.138)       0:00:20.949 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.143)       0:00:21.093 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.154)       0:00:21.247 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.151)       0:00:21.399 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.155)       0:00:21.554 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nThursday 26 March 2026  02:47:29 +0000 (0:00:00.154)       0:00:21.709 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nThursday 26 March 2026  02:47:30 +0000 (0:00:00.144)       0:00:21.854 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nThursday 26 March 2026  02:47:30 +0000 (0:00:00.370)       0:00:22.224 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nThursday 26 March 2026  02:47:30 +0000 (0:00:00.149)       0:00:22.373 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:47:30 +0000 (0:00:00.162)       0:00:22.535 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nThursday 26 March 2026  02:47:30 +0000 (0:00:00.156)       0:00:22.692 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.149)       0:00:22.842 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.150)       0:00:22.992 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.142)       0:00:23.134 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.170)       0:00:23.305 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.167)       0:00:23.472 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nThursday 26 March 2026  02:47:31 +0000 (0:00:00.166)       0:00:23.639 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nThursday 26 March 2026  02:47:32 +0000 (0:00:00.166)       0:00:23.806 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nThursday 26 March 2026  02:47:32 +0000 (0:00:00.415)       0:00:24.221 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nThursday 26 March 2026  02:47:32 +0000 (0:00:00.171)       0:00:24.392 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nThursday 26 March 2026  02:47:32 +0000 (0:00:00.180)       0:00:24.573 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nThursday 26 March 2026  02:47:33 +0000 (0:00:00.184)       0:00:24.757 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nThursday 26 March 2026  02:47:33 +0000 (0:00:00.531)       0:00:25.289 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nThursday 26 March 2026  02:47:34 +0000 (0:00:00.530)       0:00:25.819 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nThursday 26 March 2026  02:47:34 +0000 (0:00:00.165)       0:00:25.985 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'vg_name': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'vg_name': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nThursday 26 March 2026  02:47:34 +0000 (0:00:00.211)       0:00:26.197 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nThursday 26 March 2026  02:47:34 +0000 (0:00:00.180)       0:00:26.377 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nThursday 26 March 2026  02:47:34 +0000 (0:00:00.186)       0:00:26.564 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nThursday 26 March 2026  02:47:35 +0000 (0:00:00.190)       0:00:26.754 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-e2623153-bc41-510f-8884-ef957bb96082\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:47:35 +0000 (0:00:00.577)       0:00:27.332 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:47:35 +0000 (0:00:00.281)       0:00:27.613 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:36 +0000 (0:00:00.254)       0:00:27.868 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:36 +0000 (0:00:00.487)       0:00:28.355 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:36 +0000 (0:00:00.219)       0:00:28.575 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:37 +0000 (0:00:00.232)       0:00:28.807 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:37 +0000 (0:00:00.214)       0:00:29.022 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:37 +0000 (0:00:00.206)       0:00:29.228 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:37 +0000 (0:00:00.229)       0:00:29.457 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:38 +0000 (0:00:00.689)       0:00:30.147 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:38 +0000 (0:00:00.240)       0:00:30.387 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:38 +0000 (0:00:00.218)       0:00:30.606 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:39 +0000 (0:00:00.445)       0:00:31.051 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:39 +0000 (0:00:00.482)       0:00:31.534 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:40 +0000 (0:00:00.486)       0:00:32.021 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:47:40 +0000 (0:00:00.503)       0:00:32.524 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:41 +0000 (0:00:00.376)       0:00:32.900 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:41 +0000 (0:00:00.439)       0:00:33.340 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:41 +0000 (0:00:00.210)       0:00:33.551 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:42 +0000 (0:00:00.248)       0:00:33.799 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:42 +0000 (0:00:00.706)       0:00:34.506 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:42 +0000 (0:00:00.221)       0:00:34.727 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:43 +0000 (0:00:00.239)       0:00:34.966 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:43 +0000 (0:00:00.215)       0:00:35.182 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:43 +0000 (0:00:00.229)       0:00:35.412 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:43 +0000 (0:00:00.230)       0:00:35.642 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:44 +0000 (0:00:00.728)       0:00:36.371 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:44 +0000 (0:00:00.236)       0:00:36.608 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:45 +0000 (0:00:00.222)       0:00:36.831 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:47:45 +0000 (0:00:00.241)       0:00:37.073 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nThursday 26 March 2026  02:47:45 +0000 (0:00:00.219)       0:00:37.292 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nThursday 26 March 2026  02:47:45 +0000 (0:00:00.384)       0:00:37.677 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nThursday 26 March 2026  02:47:46 +0000 (0:00:00.206)       0:00:37.883 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nThursday 26 March 2026  02:47:47 +0000 (0:00:01.739)       0:00:39.622 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nThursday 26 March 2026  02:47:48 +0000 (0:00:00.244)       0:00:39.867 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nThursday 26 March 2026  02:47:49 +0000 (0:00:01.289)       0:00:41.156 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nThursday 26 March 2026  02:47:49 +0000 (0:00:00.187)       0:00:41.344 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nThursday 26 March 2026  02:47:49 +0000 (0:00:00.157)       0:00:41.501 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nThursday 26 March 2026  02:47:49 +0000 (0:00:00.183)       0:00:41.685 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nThursday 26 March 2026  02:47:50 +0000 (0:00:00.149)       0:00:41.835 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nThursday 26 March 2026  02:47:50 +0000 (0:00:00.159)       0:00:41.995 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nThursday 26 March 2026  02:47:50 +0000 (0:00:00.169)       0:00:42.164 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nThursday 26 March 2026  02:47:50 +0000 (0:00:00.165)       0:00:42.330 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nThursday 26 March 2026  02:47:50 +0000 (0:00:00.197)       0:00:42.528 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nThursday 26 March 2026  02:47:51 +0000 (0:00:00.448)       0:00:42.976 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nThursday 26 March 2026  02:47:51 +0000 (0:00:00.161)       0:00:43.138 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nThursday 26 March 2026  02:47:51 +0000 (0:00:00.167)       0:00:43.306 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nThursday 26 March 2026  02:47:51 +0000 (0:00:00.172)       0:00:43.478 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nThursday 26 March 2026  02:47:51 +0000 (0:00:00.142)       0:00:43.620 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nThursday 26 March 2026  02:47:52 +0000 (0:00:00.155)       0:00:43.776 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nThursday 26 March 2026  02:47:52 +0000 (0:00:00.141)       0:00:43.917 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nThursday 26 March 2026  02:47:52 +0000 (0:00:00.147)       0:00:44.065 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nThursday 26 March 2026  02:47:52 +0000 (0:00:00.157)       0:00:44.223 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nThursday 26 March 2026  02:47:53 +0000 (0:00:00.520)       0:00:44.743 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nThursday 26 March 2026  02:47:53 +0000 (0:00:00.497)       0:00:45.241 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nThursday 26 March 2026  02:47:53 +0000 (0:00:00.496)       0:00:45.738 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nThursday 26 March 2026  02:47:54 +0000 (0:00:00.418)       0:00:46.157 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nThursday 26 March 2026  02:47:54 +0000 (0:00:00.124)       0:00:46.281 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nThursday 26 March 2026  02:47:54 +0000 (0:00:00.133)       0:00:46.415 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nThursday 26 March 2026  02:47:54 +0000 (0:00:00.159)       0:00:46.575 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nThursday 26 March 2026  02:47:54 +0000 (0:00:00.154)       0:00:46.729 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.144)       0:00:46.873 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.171)       0:00:47.045 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.145)       0:00:47.191 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.143)       0:00:47.334 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.150)       0:00:47.484 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nThursday 26 March 2026  02:47:55 +0000 (0:00:00.139)       0:00:47.624 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nThursday 26 March 2026  02:47:56 +0000 (0:00:00.156)       0:00:47.781 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nThursday 26 March 2026  02:47:56 +0000 (0:00:00.375)       0:00:48.156 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:47:56 +0000 (0:00:00.148)       0:00:48.305 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nThursday 26 March 2026  02:47:56 +0000 (0:00:00.207)       0:00:48.513 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nThursday 26 March 2026  02:47:56 +0000 (0:00:00.156)       0:00:48.669 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.171)       0:00:48.841 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.150)       0:00:48.991 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.213)       0:00:49.205 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.183)       0:00:49.388 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.160)       0:00:49.549 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nThursday 26 March 2026  02:47:57 +0000 (0:00:00.153)       0:00:49.703 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nThursday 26 March 2026  02:47:58 +0000 (0:00:00.174)       0:00:49.877 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nThursday 26 March 2026  02:47:58 +0000 (0:00:00.156)       0:00:50.034 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nThursday 26 March 2026  02:47:58 +0000 (0:00:00.413)       0:00:50.448 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nThursday 26 March 2026  02:47:58 +0000 (0:00:00.182)       0:00:50.631 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nThursday 26 March 2026  02:47:59 +0000 (0:00:00.498)       0:00:51.129 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nThursday 26 March 2026  02:47:59 +0000 (0:00:00.563)       0:00:51.692 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nThursday 26 March 2026  02:48:00 +0000 (0:00:00.177)       0:00:51.870 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'vg_name': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'vg_name': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nThursday 26 March 2026  02:48:00 +0000 (0:00:00.169)       0:00:52.039 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nThursday 26 March 2026  02:48:00 +0000 (0:00:00.166)       0:00:52.206 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nThursday 26 March 2026  02:48:00 +0000 (0:00:00.164)       0:00:52.370 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nThursday 26 March 2026  02:48:00 +0000 (0:00:00.169)       0:00:52.539 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-a652979e-9f40-503a-bbc8-6de5e605991e\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  02:48:01 +0000 (0:00:00.319)       0:00:52.858 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nThursday 26 March 2026  02:48:01 +0000 (0:00:00.763)       0:00:53.622 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:02 +0000 (0:00:00.261)       0:00:53.883 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:02 +0000 (0:00:00.428)       0:00:54.311 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:02 +0000 (0:00:00.209)       0:00:54.521 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:03 +0000 (0:00:00.238)       0:00:54.759 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:03 +0000 (0:00:00.208)       0:00:54.968 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:03 +0000 (0:00:00.231)       0:00:55.199 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:03 +0000 (0:00:00.225)       0:00:55.425 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:03 +0000 (0:00:00.221)       0:00:55.646 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:04 +0000 (0:00:00.238)       0:00:55.885 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:04 +0000 (0:00:00.205)       0:00:56.090 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:05 +0000 (0:00:00.934)       0:00:57.024 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:05 +0000 (0:00:00.451)       0:00:57.476 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:06 +0000 (0:00:00.474)       0:00:57.951 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nThursday 26 March 2026  02:48:06 +0000 (0:00:00.494)       0:00:58.445 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:07 +0000 (0:00:00.362)       0:00:58.808 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:07 +0000 (0:00:00.454)       0:00:59.262 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:07 +0000 (0:00:00.222)       0:00:59.485 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:07 +0000 (0:00:00.210)       0:00:59.695 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:08 +0000 (0:00:00.223)       0:00:59.919 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:08 +0000 (0:00:00.241)       0:01:00.160 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:09 +0000 (0:00:00.699)       0:01:00.860 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:09 +0000 (0:00:00.234)       0:01:01.094 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:09 +0000 (0:00:00.219)       0:01:01.314 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:09 +0000 (0:00:00.269)       0:01:01.583 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:10 +0000 (0:00:00.720)       0:01:02.304 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:10 +0000 (0:00:00.227)       0:01:02.531 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:11 +0000 (0:00:00.217)       0:01:02.749 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nThursday 26 March 2026  02:48:11 +0000 (0:00:00.233)       0:01:02.983 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nThursday 26 March 2026  02:48:11 +0000 (0:00:00.221)       0:01:03.205 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nThursday 26 March 2026  02:48:11 +0000 (0:00:00.145)       0:01:03.350 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nThursday 26 March 2026  02:48:11 +0000 (0:00:00.178)       0:01:03.529 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nThursday 26 March 2026  02:48:13 +0000 (0:00:01.905)       0:01:05.434 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nThursday 26 March 2026  02:48:14 +0000 (0:00:00.418)       0:01:05.852 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nThursday 26 March 2026  02:48:15 +0000 (0:00:01.377)       0:01:07.230 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nThursday 26 March 2026  02:48:15 +0000 (0:00:00.180)       0:01:07.410 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nThursday 26 March 2026  02:48:15 +0000 (0:00:00.166)       0:01:07.576 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.202)       0:01:07.779 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.166)       0:01:07.945 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.163)       0:01:08.109 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.153)       0:01:08.262 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.161)       0:01:08.424 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nThursday 26 March 2026  02:48:16 +0000 (0:00:00.153)       0:01:08.577 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nThursday 26 March 2026  02:48:17 +0000 (0:00:00.170)       0:01:08.747 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nThursday 26 March 2026  02:48:17 +0000 (0:00:00.183)       0:01:08.931 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nThursday 26 March 2026  02:48:17 +0000 (0:00:00.416)       0:01:09.348 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nThursday 26 March 2026  02:48:17 +0000 (0:00:00.155)       0:01:09.504 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nThursday 26 March 2026  02:48:17 +0000 (0:00:00.145)       0:01:09.649 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nThursday 26 March 2026  02:48:18 +0000 (0:00:00.155)       0:01:09.804 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nThursday 26 March 2026  02:48:18 +0000 (0:00:00.224)       0:01:10.029 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nThursday 26 March 2026  02:48:18 +0000 (0:00:00.176)       0:01:10.206 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nThursday 26 March 2026  02:48:18 +0000 (0:00:00.148)       0:01:10.354 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nThursday 26 March 2026  02:48:19 +0000 (0:00:00.528)       0:01:10.883 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nThursday 26 March 2026  02:48:19 +0000 (0:00:00.541)       0:01:11.424 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nThursday 26 March 2026  02:48:20 +0000 (0:00:00.518)       0:01:11.943 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nThursday 26 March 2026  02:48:20 +0000 (0:00:00.157)       0:01:12.100 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nThursday 26 March 2026  02:48:20 +0000 (0:00:00.131)       0:01:12.232 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nThursday 26 March 2026  02:48:20 +0000 (0:00:00.384)       0:01:12.616 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.164)       0:01:12.780 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.173)       0:01:12.953 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.156)       0:01:13.110 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.138)       0:01:13.249 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.171)       0:01:13.421 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.166)       0:01:13.587 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nThursday 26 March 2026  02:48:21 +0000 (0:00:00.127)       0:01:13.714 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nThursday 26 March 2026  02:48:22 +0000 (0:00:00.152)       0:01:13.867 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nThursday 26 March 2026  02:48:22 +0000 (0:00:00.136)       0:01:14.003 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nThursday 26 March 2026  02:48:22 +0000 (0:00:00.151)       0:01:14.154 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:48:22 +0000 (0:00:00.157)       0:01:14.312 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nThursday 26 March 2026  02:48:22 +0000 (0:00:00.390)       0:01:14.702 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.158)       0:01:14.861 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.160)       0:01:15.021 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.153)       0:01:15.175 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.174)       0:01:15.349 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.184)       0:01:15.534 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nThursday 26 March 2026  02:48:23 +0000 (0:00:00.172)       0:01:15.706 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nThursday 26 March 2026  02:48:24 +0000 (0:00:00.191)       0:01:15.898 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nThursday 26 March 2026  02:48:24 +0000 (0:00:00.179)       0:01:16.077 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nThursday 26 March 2026  02:48:24 +0000 (0:00:00.162)       0:01:16.240 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nThursday 26 March 2026  02:48:24 +0000 (0:00:00.199)       0:01:16.440 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nThursday 26 March 2026  02:48:24 +0000 (0:00:00.167)       0:01:16.607 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nThursday 26 March 2026  02:48:25 +0000 (0:00:00.744)       0:01:17.352 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nThursday 26 March 2026  02:48:26 +0000 (0:00:00.515)       0:01:17.868 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nThursday 26 March 2026  02:48:26 +0000 (0:00:00.185)       0:01:18.054 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'vg_name': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'vg_name': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nThursday 26 March 2026  02:48:26 +0000 (0:00:00.185)       0:01:18.239 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nThursday 26 March 2026  02:48:26 +0000 (0:00:00.187)       0:01:18.427 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nThursday 26 March 2026  02:48:26 +0000 (0:00:00.175)       0:01:18.602 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nThursday 26 March 2026  02:48:27 +0000 (0:00:00.184)       0:01:18.786 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:48:27 +0000 (0:00:00.154)       0:01:18.940 ******** \n=============================================================================== \nCreate block VGs -------------------------------------------------------- 5.70s\nCreate block LVs -------------------------------------------------------- 4.18s\nGet list of Ceph LVs with associated VGs -------------------------------- 1.77s\nGather DB VGs with total and available size in bytes -------------------- 1.75s\nGet list of Ceph PVs with associated VGs -------------------------------- 1.61s\nGather WAL VGs with total and available size in bytes ------------------- 1.54s\nGather DB+WAL VGs with total and available size in bytes ---------------- 1.54s\nAdd known links to the list of available block devices ------------------ 1.51s\nAdd known partitions to the list of available block devices ------------- 1.36s\nGet extra vars for Ceph configuration ----------------------------------- 1.31s\nPrint LVM report data --------------------------------------------------- 1.05s\nPrint 'Create block VGs' ------------------------------------------------ 1.05s\nAdd known links to the list of available block devices ------------------ 0.98s\nCombine JSON from _db/wal/db_wal_vgs_cmd_output ------------------------- 0.97s\nAdd known links to the list of available block devices ------------------ 0.93s\nGet initial list of available block devices ----------------------------- 0.82s\nCount OSDs put on ceph_db_wal_devices defined in lvm_volumes ------------ 0.82s\nCreate DB LVs for ceph_db_wal_devices ----------------------------------- 0.79s\nCount OSDs put on ceph_db_devices defined in lvm_volumes ---------------- 0.79s\nCreate WAL LVs for ceph_db_wal_devices ---------------------------------- 0.77s\n\u001b[32m2026-03-26 02:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1dc7d21e-cb25-4027-b145-729ca3421def (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1dc7d21e-cb25-4027-b145-729ca3421def (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nThursday 26 March 2026  02:48:44 +0000 (0:00:00.280)       0:00:00.280 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nThursday 26 March 2026  02:48:45 +0000 (0:00:01.150)       0:00:01.431 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  02:48:47 +0000 (0:00:01.369)       0:00:02.800 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nThursday 26 March 2026  02:48:52 +0000 (0:00:04.942)       0:00:07.742 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  02:48:52 +0000 (0:00:00.595)       0:00:08.338 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.94s\nosism.commons.facts : Copy fact files ----------------------------------- 1.37s\nosism.commons.facts : Create custom facts directory --------------------- 1.15s\nGather facts for all hosts ---------------------------------------------- 0.60s\n\u001b[32m2026-03-26 02:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0afd95ab-3afe-4cf0-a623-ba200f82b57e (ceph) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 02:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0afd95ab-3afe-4cf0-a623-ba200f82b57e (ceph) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Prepare deployment of Ceph services] *************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  02:49:01 +0000 (0:00:00.861)       0:00:00.861 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  02:49:02 +0000 (0:00:01.228)       0:00:02.090 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  02:49:03 +0000 (0:00:01.368)       0:00:03.459 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  02:49:04 +0000 (0:00:01.039)       0:00:04.498 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  02:49:06 +0000 (0:00:01.044)       0:00:05.542 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  02:49:07 +0000 (0:00:01.090)       0:00:06.633 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  02:49:07 +0000 (0:00:00.670)       0:00:07.304 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  02:49:08 +0000 (0:00:00.857)       0:00:08.161 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  02:49:09 +0000 (0:00:00.633)       0:00:08.795 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  02:49:10 +0000 (0:00:00.826)       0:00:09.621 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  02:49:10 +0000 (0:00:00.714)       0:00:10.336 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  02:49:11 +0000 (0:00:00.712)       0:00:11.048 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  02:49:14 +0000 (0:00:02.505)       0:00:13.554 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  02:49:14 +0000 (0:00:00.438)       0:00:13.992 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  02:49:15 +0000 (0:00:00.653)       0:00:14.646 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  02:49:15 +0000 (0:00:00.169)       0:00:14.816 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 02:49:12.428284', 'end': '2026-03-26 02:49:12.465831', 'delta': '0:00:00.037547', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 02:49:13.089373', 'end': '2026-03-26 02:49:13.140350', 'delta': '0:00:00.050977', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 02:49:13.625707', 'end': '2026-03-26 02:49:13.679965', 'delta': '0:00:00.054258', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  02:49:15 +0000 (0:00:00.270)       0:00:15.086 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  02:49:16 +0000 (0:00:00.749)       0:00:15.836 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  02:49:17 +0000 (0:00:01.088)       0:00:16.925 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  02:49:18 +0000 (0:00:00.632)       0:00:17.558 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  02:49:19 +0000 (0:00:01.181)       0:00:18.739 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  02:49:19 +0000 (0:00:00.710)       0:00:19.450 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  02:49:20 +0000 (0:00:00.145)       0:00:19.595 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  02:49:20 +0000 (0:00:00.234)       0:00:19.829 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  02:49:21 +0000 (0:00:00.828)       0:00:20.658 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  02:49:21 +0000 (0:00:00.666)       0:00:21.325 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  02:49:22 +0000 (0:00:00.836)       0:00:22.162 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  02:49:23 +0000 (0:00:00.625)       0:00:22.787 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  02:49:24 +0000 (0:00:00.880)       0:00:23.668 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  02:49:24 +0000 (0:00:00.647)       0:00:24.315 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  02:49:25 +0000 (0:00:00.917)       0:00:25.233 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-12-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-16-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  02:49:26 +0000 (0:00:01.059)       0:00:26.292 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-12-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-16-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  02:49:28 +0000 (0:00:01.520)       0:00:27.813 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  02:49:29 +0000 (0:00:00.948)       0:00:28.761 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  02:49:30 +0000 (0:00:00.885)       0:00:29.647 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  02:49:31 +0000 (0:00:00.928)       0:00:30.575 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  02:49:31 +0000 (0:00:00.926)       0:00:31.501 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  02:49:32 +0000 (0:00:00.664)       0:00:32.166 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  02:49:33 +0000 (0:00:00.925)       0:00:33.092 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  02:49:35 +0000 (0:00:01.820)       0:00:34.912 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  02:49:36 +0000 (0:00:01.050)       0:00:35.963 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  02:49:37 +0000 (0:00:01.128)       0:00:37.091 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  02:49:37 +0000 (0:00:00.347)       0:00:37.439 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  02:49:38 +0000 (0:00:00.379)       0:00:37.819 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  02:49:38 +0000 (0:00:00.534)       0:00:38.354 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  02:49:39 +0000 (0:00:00.460)       0:00:38.815 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  02:49:39 +0000 (0:00:00.387)       0:00:39.203 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  02:49:40 +0000 (0:00:00.428)       0:00:39.631 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  02:49:40 +0000 (0:00:00.401)       0:00:40.032 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  02:49:40 +0000 (0:00:00.355)       0:00:40.388 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  02:49:41 +0000 (0:00:01.045)       0:00:41.433 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  02:49:42 +0000 (0:00:00.821)       0:00:42.255 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:49:44 +0000 (0:00:02.014)       0:00:44.269 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:49:46 +0000 (0:00:01.312)       0:00:45.582 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:49:47 +0000 (0:00:01.295)       0:00:46.877 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:49:48 +0000 (0:00:01.274)       0:00:48.152 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:49:49 +0000 (0:00:00.747)       0:00:48.900 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:49:50 +0000 (0:00:01.056)       0:00:49.957 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:49:51 +0000 (0:00:00.725)       0:00:50.682 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:49:52 +0000 (0:00:01.363)       0:00:52.046 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:49:53 +0000 (0:00:00.664)       0:00:52.710 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:49:54 +0000 (0:00:00.893)       0:00:53.604 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:49:55 +0000 (0:00:01.075)       0:00:54.679 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:49:56 +0000 (0:00:01.406)       0:00:56.086 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:49:57 +0000 (0:00:00.657)       0:00:56.744 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:49:58 +0000 (0:00:00.990)       0:00:57.735 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:49:58 +0000 (0:00:00.682)       0:00:58.417 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:49:59 +0000 (0:00:00.876)       0:00:59.293 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:50:00 +0000 (0:00:00.609)       0:00:59.903 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:50:01 +0000 (0:00:00.898)       0:01:00.801 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:50:01 +0000 (0:00:00.603)       0:01:01.405 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:50:02 +0000 (0:00:00.919)       0:01:02.324 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:50:03 +0000 (0:00:00.998)       0:01:03.322 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  02:50:05 +0000 (0:00:01.382)       0:01:04.705 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  02:50:06 +0000 (0:00:01.512)       0:01:06.217 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  02:50:08 +0000 (0:00:02.252)       0:01:08.470 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  02:50:10 +0000 (0:00:01.342)       0:01:09.812 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  02:50:10 +0000 (0:00:00.682)       0:01:10.495 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  02:50:11 +0000 (0:00:00.966)       0:01:11.462 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  02:50:13 +0000 (0:00:01.348)       0:01:12.810 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  02:50:14 +0000 (0:00:01.234)       0:01:14.045 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  02:50:15 +0000 (0:00:00.679)       0:01:14.725 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  02:50:16 +0000 (0:00:00.941)       0:01:15.667 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  02:50:16 +0000 (0:00:00.626)       0:01:16.293 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  02:50:18 +0000 (0:00:01.326)       0:01:17.620 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  02:51:22 +0000 (0:01:04.165)       0:02:21.785 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  02:51:23 +0000 (0:00:00.742)       0:02:22.528 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  02:51:23 +0000 (0:00:00.947)       0:02:23.475 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  02:51:24 +0000 (0:00:00.167)       0:02:23.643 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  02:51:24 +0000 (0:00:00.674)       0:02:24.318 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  02:51:25 +0000 (0:00:00.881)       0:02:25.199 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  02:51:26 +0000 (0:00:00.695)       0:02:25.894 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  02:51:29 +0000 (0:00:03.470)       0:02:29.365 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  02:51:30 +0000 (0:00:00.685)       0:02:30.050 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  02:51:31 +0000 (0:00:01.322)       0:02:31.372 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  02:51:32 +0000 (0:00:00.904)       0:02:32.276 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  02:51:33 +0000 (0:00:00.626)       0:02:32.903 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  02:51:34 +0000 (0:00:00.898)       0:02:33.801 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  02:51:34 +0000 (0:00:00.702)       0:02:34.503 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  02:51:35 +0000 (0:00:00.898)       0:02:35.402 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  02:51:36 +0000 (0:00:00.664)       0:02:36.066 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  02:51:37 +0000 (0:00:00.963)       0:02:37.030 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  02:51:38 +0000 (0:00:00.915)       0:02:37.945 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  02:51:39 +0000 (0:00:01.391)       0:02:39.337 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  02:51:41 +0000 (0:00:01.321)       0:02:40.658 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  02:51:47 +0000 (0:00:06.356)       0:02:47.015 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  02:51:48 +0000 (0:00:01.133)       0:02:48.148 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  02:51:49 +0000 (0:00:00.745)       0:02:48.894 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  02:51:50 +0000 (0:00:01.240)       0:02:50.134 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  02:51:51 +0000 (0:00:00.920)       0:02:51.054 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  02:51:52 +0000 (0:00:00.665)       0:02:51.719 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  02:51:53 +0000 (0:00:00.889)       0:02:52.609 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  02:51:53 +0000 (0:00:00.631)       0:02:53.240 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  02:51:54 +0000 (0:00:00.939)       0:02:54.180 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  02:51:55 +0000 (0:00:00.688)       0:02:54.869 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  02:51:56 +0000 (0:00:00.934)       0:02:55.803 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  02:51:56 +0000 (0:00:00.685)       0:02:56.489 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  02:51:59 +0000 (0:00:02.982)       0:02:59.471 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  02:52:00 +0000 (0:00:00.688)       0:03:00.159 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  02:52:01 +0000 (0:00:00.963)       0:03:01.122 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  02:52:02 +0000 (0:00:00.895)       0:03:02.018 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  02:52:03 +0000 (0:00:00.702)       0:03:02.720 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  02:52:04 +0000 (0:00:01.098)       0:03:03.818 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  02:52:04 +0000 (0:00:00.647)       0:03:04.466 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  02:52:05 +0000 (0:00:00.871)       0:03:05.337 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  02:52:06 +0000 (0:00:00.713)       0:03:06.051 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  02:52:07 +0000 (0:00:00.860)       0:03:06.912 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  02:52:08 +0000 (0:00:00.682)       0:03:07.594 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  02:52:08 +0000 (0:00:00.894)       0:03:08.489 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  02:52:09 +0000 (0:00:00.460)       0:03:08.949 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  02:52:09 +0000 (0:00:00.447)       0:03:09.396 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  02:52:10 +0000 (0:00:00.466)       0:03:09.863 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  02:52:11 +0000 (0:00:00.670)       0:03:10.534 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  02:52:12 +0000 (0:00:01.812)       0:03:12.346 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  02:52:15 +0000 (0:00:02.709)       0:03:15.056 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  02:52:16 +0000 (0:00:01.012)       0:03:16.069 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nThursday 26 March 2026  02:52:17 +0000 (0:00:01.151)       0:03:17.220 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nThursday 26 March 2026  02:52:18 +0000 (0:00:00.329)       0:03:17.550 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nThursday 26 March 2026  02:52:19 +0000 (0:00:01.531)       0:03:19.081 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nThursday 26 March 2026  02:52:20 +0000 (0:00:00.693)       0:03:19.775 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  02:52:20 +0000 (0:00:00.363)       0:03:20.139 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nThursday 26 March 2026  02:52:21 +0000 (0:00:01.200)       0:03:21.340 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nThursday 26 March 2026  02:52:22 +0000 (0:00:00.429)       0:03:21.769 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nThursday 26 March 2026  02:52:22 +0000 (0:00:00.336)       0:03:22.105 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nThursday 26 March 2026  02:52:22 +0000 (0:00:00.246)       0:03:22.352 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nThursday 26 March 2026  02:52:23 +0000 (0:00:00.546)       0:03:22.899 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nThursday 26 March 2026  02:52:23 +0000 (0:00:00.261)       0:03:23.160 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nThursday 26 March 2026  02:52:23 +0000 (0:00:00.255)       0:03:23.416 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nThursday 26 March 2026  02:52:24 +0000 (0:00:00.150)       0:03:23.567 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nThursday 26 March 2026  02:52:24 +0000 (0:00:00.259)       0:03:23.827 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nThursday 26 March 2026  02:52:24 +0000 (0:00:00.238)       0:03:24.065 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nThursday 26 March 2026  02:52:25 +0000 (0:00:00.455)       0:03:24.520 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nThursday 26 March 2026  02:52:25 +0000 (0:00:00.331)       0:03:24.852 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nThursday 26 March 2026  02:52:25 +0000 (0:00:00.265)       0:03:25.117 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  02:52:26 +0000 (0:00:00.782)       0:03:25.900 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nThursday 26 March 2026  02:52:27 +0000 (0:00:00.893)       0:03:26.794 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nThursday 26 March 2026  02:52:27 +0000 (0:00:00.562)       0:03:27.356 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nThursday 26 March 2026  02:52:29 +0000 (0:00:01.284)       0:03:28.641 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nThursday 26 March 2026  02:52:29 +0000 (0:00:00.679)       0:03:29.321 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  02:52:30 +0000 (0:00:00.331)       0:03:29.652 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nThursday 26 March 2026  02:52:31 +0000 (0:00:01.196)       0:03:30.849 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nThursday 26 March 2026  02:52:31 +0000 (0:00:00.369)       0:03:31.219 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nThursday 26 March 2026  02:52:32 +0000 (0:00:01.245)       0:03:32.465 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nThursday 26 March 2026  02:52:34 +0000 (0:00:01.165)       0:03:33.630 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  02:52:34 +0000 (0:00:00.386)       0:03:34.017 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  02:52:35 +0000 (0:00:00.727)       0:03:34.744 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nThursday 26 March 2026  02:52:36 +0000 (0:00:01.149)       0:03:35.894 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nThursday 26 March 2026  02:52:36 +0000 (0:00:00.381)       0:03:36.275 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nThursday 26 March 2026  02:52:38 +0000 (0:00:01.464)       0:03:37.740 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nThursday 26 March 2026  02:52:38 +0000 (0:00:00.659)       0:03:38.399 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mon] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:52:39 +0000 (0:00:00.630)       0:03:39.030 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:52:40 +0000 (0:00:00.811)       0:03:39.841 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:52:40 +0000 (0:00:00.567)       0:03:40.409 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:52:41 +0000 (0:00:00.741)       0:03:41.151 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:52:42 +0000 (0:00:00.587)       0:03:41.738 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:52:42 +0000 (0:00:00.375)       0:03:42.114 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:52:42 +0000 (0:00:00.317)       0:03:42.431 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:52:43 +0000 (0:00:00.729)       0:03:43.160 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:52:44 +0000 (0:00:00.626)       0:03:43.787 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:52:44 +0000 (0:00:00.399)       0:03:44.187 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:52:45 +0000 (0:00:00.853)       0:03:45.040 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:52:46 +0000 (0:00:00.753)       0:03:45.793 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:52:46 +0000 (0:00:00.609)       0:03:46.403 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:52:47 +0000 (0:00:00.373)       0:03:46.776 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:52:47 +0000 (0:00:00.324)       0:03:47.101 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:52:48 +0000 (0:00:00.576)       0:03:47.677 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:52:48 +0000 (0:00:00.356)       0:03:48.033 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:52:48 +0000 (0:00:00.355)       0:03:48.389 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:52:49 +0000 (0:00:00.348)       0:03:48.737 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:52:49 +0000 (0:00:00.666)       0:03:49.404 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:52:50 +0000 (0:00:00.379)       0:03:49.784 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  02:52:50 +0000 (0:00:00.593)       0:03:50.377 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nThursday 26 March 2026  02:52:51 +0000 (0:00:00.624)       0:03:51.002 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nThursday 26 March 2026  02:52:52 +0000 (0:00:00.610)       0:03:51.613 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nThursday 26 March 2026  02:52:52 +0000 (0:00:00.156)       0:03:51.769 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nThursday 26 March 2026  02:52:53 +0000 (0:00:01.067)       0:03:52.836 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nThursday 26 March 2026  02:52:53 +0000 (0:00:00.613)       0:03:53.450 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nThursday 26 March 2026  02:52:54 +0000 (0:00:00.378)       0:03:53.828 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nThursday 26 March 2026  02:52:55 +0000 (0:00:01.161)       0:03:54.990 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nThursday 26 March 2026  02:52:56 +0000 (0:00:00.830)       0:03:55.820 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nThursday 26 March 2026  02:52:57 +0000 (0:00:00.950)       0:03:56.771 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nThursday 26 March 2026  02:52:57 +0000 (0:00:00.730)       0:03:57.501 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nThursday 26 March 2026  02:52:59 +0000 (0:00:01.250)       0:03:58.752 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nThursday 26 March 2026  02:52:59 +0000 (0:00:00.730)       0:03:59.483 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nThursday 26 March 2026  02:53:03 +0000 (0:00:03.125)       0:04:02.608 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nThursday 26 March 2026  02:53:04 +0000 (0:00:01.266)       0:04:03.874 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nThursday 26 March 2026  02:53:05 +0000 (0:00:00.657)       0:04:04.532 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nThursday 26 March 2026  02:53:05 +0000 (0:00:00.355)       0:04:04.887 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nThursday 26 March 2026  02:53:06 +0000 (0:00:01.533)       0:04:06.420 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nThursday 26 March 2026  02:53:08 +0000 (0:00:01.632)       0:04:08.052 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nThursday 26 March 2026  02:53:08 +0000 (0:00:00.334)       0:04:08.387 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  02:53:09 +0000 (0:00:00.567)       0:04:08.954 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nThursday 26 March 2026  02:53:10 +0000 (0:00:00.631)       0:04:09.585 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  02:53:10 +0000 (0:00:00.350)       0:04:09.936 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nThursday 26 March 2026  02:53:11 +0000 (0:00:00.590)       0:04:10.527 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nThursday 26 March 2026  02:53:13 +0000 (0:00:02.129)       0:04:12.656 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nThursday 26 March 2026  02:53:14 +0000 (0:00:01.196)       0:04:13.852 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nThursday 26 March 2026  02:53:16 +0000 (0:00:01.859)       0:04:15.712 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nThursday 26 March 2026  02:53:18 +0000 (0:00:01.955)       0:04:17.667 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nThursday 26 March 2026  02:53:19 +0000 (0:00:00.897)       0:04:18.565 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nThursday 26 March 2026  02:53:40 +0000 (0:00:21.851)       0:04:40.416 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nThursday 26 March 2026  02:53:49 +0000 (0:00:08.986)       0:04:49.402 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nThursday 26 March 2026  02:53:50 +0000 (0:00:00.363)       0:04:49.766 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}]) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  02:54:05 +0000 (0:00:14.915)       0:05:04.681 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  02:54:05 +0000 (0:00:00.452)       0:05:05.134 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nThursday 26 March 2026  02:54:06 +0000 (0:00:00.870)       0:05:06.004 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nThursday 26 March 2026  02:54:06 +0000 (0:00:00.357)       0:05:06.362 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nThursday 26 March 2026  02:54:07 +0000 (0:00:00.385)       0:05:06.748 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nThursday 26 March 2026  02:54:08 +0000 (0:00:00.959)       0:05:07.707 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mgr] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:54:09 +0000 (0:00:00.899)       0:05:08.607 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:54:09 +0000 (0:00:00.580)       0:05:09.188 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:54:10 +0000 (0:00:00.840)       0:05:10.028 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:54:11 +0000 (0:00:00.768)       0:05:10.796 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:54:11 +0000 (0:00:00.352)       0:05:11.148 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:54:12 +0000 (0:00:00.577)       0:05:11.726 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:54:12 +0000 (0:00:00.354)       0:05:12.081 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:54:13 +0000 (0:00:00.788)       0:05:12.870 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:54:13 +0000 (0:00:00.335)       0:05:13.205 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:54:14 +0000 (0:00:00.644)       0:05:13.850 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:54:15 +0000 (0:00:00.770)       0:05:14.621 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:54:15 +0000 (0:00:00.733)       0:05:15.354 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:54:16 +0000 (0:00:00.328)       0:05:15.683 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:54:16 +0000 (0:00:00.665)       0:05:16.348 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:54:17 +0000 (0:00:00.390)       0:05:16.739 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:54:17 +0000 (0:00:00.336)       0:05:17.075 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:54:18 +0000 (0:00:00.613)       0:05:17.689 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:54:18 +0000 (0:00:00.324)       0:05:18.014 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:54:18 +0000 (0:00:00.341)       0:05:18.355 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:54:19 +0000 (0:00:00.379)       0:05:18.735 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:54:19 +0000 (0:00:00.632)       0:05:19.368 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  02:54:20 +0000 (0:00:00.601)       0:05:19.969 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  02:54:21 +0000 (0:00:00.911)       0:05:20.881 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nThursday 26 March 2026  02:54:22 +0000 (0:00:00.805)       0:05:21.686 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nThursday 26 March 2026  02:54:22 +0000 (0:00:00.742)       0:05:22.429 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nThursday 26 March 2026  02:54:23 +0000 (0:00:00.355)       0:05:22.784 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nThursday 26 March 2026  02:54:33 +0000 (0:00:10.631)       0:05:33.415 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nThursday 26 March 2026  02:54:34 +0000 (0:00:00.641)       0:05:34.056 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  02:54:36 +0000 (0:00:02.303)       0:05:36.360 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nThursday 26 March 2026  02:54:38 +0000 (0:00:01.397)       0:05:37.758 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nThursday 26 March 2026  02:54:39 +0000 (0:00:00.819)       0:05:38.577 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  02:54:39 +0000 (0:00:00.606)       0:05:39.184 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  02:54:40 +0000 (0:00:00.359)       0:05:39.544 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  02:54:40 +0000 (0:00:00.604)       0:05:40.149 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nThursday 26 March 2026  02:54:41 +0000 (0:00:00.608)       0:05:40.757 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  02:54:41 +0000 (0:00:00.367)       0:05:41.125 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nThursday 26 March 2026  02:54:42 +0000 (0:00:00.548)       0:05:41.674 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nThursday 26 March 2026  02:54:44 +0000 (0:00:01.999)       0:05:43.673 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nThursday 26 March 2026  02:54:45 +0000 (0:00:01.225)       0:05:44.898 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nThursday 26 March 2026  02:54:47 +0000 (0:00:01.786)       0:05:46.684 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  02:54:49 +0000 (0:00:02.736)       0:05:49.420 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nThursday 26 March 2026  02:54:50 +0000 (0:00:00.728)       0:05:50.149 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nThursday 26 March 2026  02:55:14 +0000 (0:00:24.280)       0:06:14.429 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nThursday 26 March 2026  02:55:16 +0000 (0:00:01.301)       0:06:15.731 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nThursday 26 March 2026  02:55:16 +0000 (0:00:00.330)       0:06:16.061 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nThursday 26 March 2026  02:55:16 +0000 (0:00:00.174)       0:06:16.236 ******** \n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nThursday 26 March 2026  02:55:23 +0000 (0:00:06.365)       0:06:22.602 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  02:55:28 +0000 (0:00:05.322)       0:06:27.925 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  02:55:29 +0000 (0:00:00.706)       0:06:28.631 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nThursday 26 March 2026  02:55:30 +0000 (0:00:00.903)       0:06:29.534 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nThursday 26 March 2026  02:55:30 +0000 (0:00:00.370)       0:06:29.905 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nThursday 26 March 2026  02:55:31 +0000 (0:00:01.162)       0:06:31.068 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nThursday 26 March 2026  02:55:32 +0000 (0:00:00.920)       0:06:31.989 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-osd] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:55:33 +0000 (0:00:00.895)       0:06:32.884 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:55:33 +0000 (0:00:00.594)       0:06:33.479 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:55:34 +0000 (0:00:00.808)       0:06:34.288 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:55:35 +0000 (0:00:00.389)       0:06:34.678 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:55:35 +0000 (0:00:00.778)       0:06:35.457 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:55:36 +0000 (0:00:00.694)       0:06:36.152 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:55:37 +0000 (0:00:01.077)       0:06:37.229 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:55:38 +0000 (0:00:00.346)       0:06:37.575 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:55:38 +0000 (0:00:00.395)       0:06:37.971 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:55:38 +0000 (0:00:00.328)       0:06:38.300 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:55:39 +0000 (0:00:01.064)       0:06:39.364 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:55:40 +0000 (0:00:00.726)       0:06:40.091 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:55:40 +0000 (0:00:00.328)       0:06:40.419 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:55:41 +0000 (0:00:00.360)       0:06:40.780 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:55:41 +0000 (0:00:00.663)       0:06:41.444 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:55:42 +0000 (0:00:00.355)       0:06:41.800 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:55:42 +0000 (0:00:00.384)       0:06:42.184 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:55:43 +0000 (0:00:00.333)       0:06:42.518 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:55:43 +0000 (0:00:00.651)       0:06:43.169 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:55:43 +0000 (0:00:00.343)       0:06:43.513 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:55:44 +0000 (0:00:00.361)       0:06:43.874 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nThursday 26 March 2026  02:55:45 +0000 (0:00:00.879)       0:06:44.754 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  02:55:45 +0000 (0:00:00.347)       0:06:45.101 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nThursday 26 March 2026  02:55:46 +0000 (0:00:00.914)       0:06:46.015 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nThursday 26 March 2026  02:55:47 +0000 (0:00:00.837)       0:06:46.853 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nThursday 26 March 2026  02:55:47 +0000 (0:00:00.338)       0:06:47.191 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nThursday 26 March 2026  02:55:48 +0000 (0:00:00.355)       0:06:47.547 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nThursday 26 March 2026  02:55:48 +0000 (0:00:00.654)       0:06:48.202 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nThursday 26 March 2026  02:55:49 +0000 (0:00:00.653)       0:06:48.855 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nThursday 26 March 2026  02:55:51 +0000 (0:00:01.974)       0:06:50.830 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nThursday 26 March 2026  02:55:51 +0000 (0:00:00.334)       0:06:51.165 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nThursday 26 March 2026  02:55:52 +0000 (0:00:00.852)       0:06:52.017 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nThursday 26 March 2026  02:55:53 +0000 (0:00:01.019)       0:06:53.037 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  02:55:55 +0000 (0:00:02.033)       0:06:55.070 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nThursday 26 March 2026  02:55:56 +0000 (0:00:01.153)       0:06:56.224 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nThursday 26 March 2026  02:55:58 +0000 (0:00:02.086)       0:06:58.311 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Use ceph-volume to create osds] *******************************\nThursday 26 March 2026  02:55:59 +0000 (0:00:00.966)       0:06:59.277 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nThursday 26 March 2026  02:56:40 +0000 (0:00:41.005)       0:07:40.283 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nThursday 26 March 2026  02:56:41 +0000 (0:00:00.358)       0:07:40.642 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nThursday 26 March 2026  02:56:41 +0000 (0:00:00.842)       0:07:41.484 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nThursday 26 March 2026  02:56:42 +0000 (0:00:00.722)       0:07:42.206 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  02:56:45 +0000 (0:00:02.698)       0:07:44.905 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nThursday 26 March 2026  02:56:46 +0000 (0:00:00.912)       0:07:45.817 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nThursday 26 March 2026  02:56:47 +0000 (0:00:01.197)       0:07:47.014 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nThursday 26 March 2026  02:56:48 +0000 (0:00:01.249)       0:07:48.264 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  02:56:50 +0000 (0:00:02.004)       0:07:50.269 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nThursday 26 March 2026  02:56:51 +0000 (0:00:00.340)       0:07:50.609 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nThursday 26 March 2026  02:56:51 +0000 (0:00:00.352)       0:07:50.962 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nThursday 26 March 2026  02:56:52 +0000 (0:00:01.029)       0:07:51.991 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nThursday 26 March 2026  02:56:54 +0000 (0:00:02.493)       0:07:54.484 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nThursday 26 March 2026  02:56:58 +0000 (0:00:03.433)       0:07:57.918 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nThursday 26 March 2026  02:57:01 +0000 (0:00:03.095)       0:08:01.014 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nThursday 26 March 2026  02:57:14 +0000 (0:00:12.703)       0:08:13.717 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  02:57:15 +0000 (0:00:01.343)       0:08:15.061 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  02:57:16 +0000 (0:00:00.684)       0:08:15.745 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nThursday 26 March 2026  02:57:16 +0000 (0:00:00.569)       0:08:16.315 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nThursday 26 March 2026  02:57:17 +0000 (0:00:00.415)       0:08:16.731 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nThursday 26 March 2026  02:57:17 +0000 (0:00:00.374)       0:08:17.105 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nThursday 26 March 2026  02:57:17 +0000 (0:00:00.252)       0:08:17.358 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nThursday 26 March 2026  02:57:18 +0000 (0:00:00.733)       0:08:18.091 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nThursday 26 March 2026  02:57:18 +0000 (0:00:00.263)       0:08:18.355 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nThursday 26 March 2026  02:57:19 +0000 (0:00:00.258)       0:08:18.614 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nThursday 26 March 2026  02:57:19 +0000 (0:00:00.163)       0:08:18.777 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nThursday 26 March 2026  02:57:19 +0000 (0:00:00.246)       0:08:19.024 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nThursday 26 March 2026  02:57:19 +0000 (0:00:00.263)       0:08:19.288 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nThursday 26 March 2026  02:57:20 +0000 (0:00:00.432)       0:08:19.720 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nThursday 26 March 2026  02:57:20 +0000 (0:00:00.640)       0:08:20.361 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nThursday 26 March 2026  02:57:21 +0000 (0:00:00.259)       0:08:20.621 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role ceph-crash] ***************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:57:21 +0000 (0:00:00.758)       0:08:21.380 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:57:23 +0000 (0:00:01.348)       0:08:22.728 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:57:24 +0000 (0:00:01.484)       0:08:24.213 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:57:26 +0000 (0:00:01.323)       0:08:25.537 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:57:26 +0000 (0:00:00.783)       0:08:26.320 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:57:27 +0000 (0:00:00.990)       0:08:27.311 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:57:28 +0000 (0:00:00.724)       0:08:28.035 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:57:29 +0000 (0:00:01.419)       0:08:29.455 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:57:30 +0000 (0:00:00.658)       0:08:30.113 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:57:31 +0000 (0:00:00.965)       0:08:31.079 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:57:32 +0000 (0:00:01.045)       0:08:32.124 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:57:34 +0000 (0:00:01.424)       0:08:33.549 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:57:34 +0000 (0:00:00.668)       0:08:34.217 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:57:35 +0000 (0:00:00.985)       0:08:35.203 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:57:36 +0000 (0:00:00.647)       0:08:35.851 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:57:37 +0000 (0:00:00.949)       0:08:36.801 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:57:37 +0000 (0:00:00.606)       0:08:37.407 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:57:38 +0000 (0:00:00.962)       0:08:38.369 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:57:39 +0000 (0:00:00.674)       0:08:39.044 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:57:40 +0000 (0:00:00.913)       0:08:39.957 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:57:41 +0000 (0:00:00.682)       0:08:40.639 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nThursday 26 March 2026  02:57:42 +0000 (0:00:01.459)       0:08:42.099 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nThursday 26 March 2026  02:57:46 +0000 (0:00:04.199)       0:08:46.299 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nThursday 26 March 2026  02:57:48 +0000 (0:00:01.847)       0:08:48.147 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nThursday 26 March 2026  02:57:50 +0000 (0:00:01.475)       0:08:49.623 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nThursday 26 March 2026  02:57:51 +0000 (0:00:01.315)       0:08:50.938 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nThursday 26 March 2026  02:57:52 +0000 (0:00:01.345)       0:08:52.284 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nThursday 26 March 2026  02:57:54 +0000 (0:00:01.481)       0:08:53.765 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************\nThursday 26 March 2026  02:57:57 +0000 (0:00:03.720)       0:08:57.486 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******\nThursday 26 March 2026  02:57:59 +0000 (0:00:01.429)       0:08:58.915 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************\nThursday 26 March 2026  02:58:00 +0000 (0:00:00.769)       0:08:59.685 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******\nThursday 26 March 2026  02:58:02 +0000 (0:00:02.440)       0:09:02.125 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mds] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:58:03 +0000 (0:00:01.349)       0:09:03.475 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:58:04 +0000 (0:00:00.584)       0:09:04.059 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:58:05 +0000 (0:00:00.894)       0:09:04.954 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:58:05 +0000 (0:00:00.345)       0:09:05.299 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:58:06 +0000 (0:00:00.765)       0:09:06.065 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:58:07 +0000 (0:00:00.668)       0:09:06.733 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:58:08 +0000 (0:00:01.029)       0:09:07.763 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:58:08 +0000 (0:00:00.348)       0:09:08.111 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:58:08 +0000 (0:00:00.337)       0:09:08.449 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:58:09 +0000 (0:00:00.331)       0:09:08.780 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:58:10 +0000 (0:00:01.088)       0:09:09.869 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:58:11 +0000 (0:00:00.723)       0:09:10.592 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:58:11 +0000 (0:00:00.355)       0:09:10.947 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:58:11 +0000 (0:00:00.362)       0:09:11.309 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:58:12 +0000 (0:00:00.651)       0:09:11.961 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:58:12 +0000 (0:00:00.372)       0:09:12.333 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:58:13 +0000 (0:00:00.373)       0:09:12.707 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:58:13 +0000 (0:00:00.353)       0:09:13.061 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:58:14 +0000 (0:00:00.616)       0:09:13.678 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:58:14 +0000 (0:00:00.338)       0:09:14.016 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:58:14 +0000 (0:00:00.352)       0:09:14.369 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nThursday 26 March 2026  02:58:15 +0000 (0:00:00.853)       0:09:15.223 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule details] *********************\nThursday 26 March 2026  02:58:16 +0000 (0:00:00.477)       0:09:15.701 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule name] ************************\nThursday 26 March 2026  02:58:18 +0000 (0:00:02.073)       0:09:17.775 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create filesystem pools] **************************************\nThursday 26 March 2026  02:58:18 +0000 (0:00:00.244)       0:09:18.020 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [ceph-mds : Create ceph filesystem] ***************************************\nThursday 26 March 2026  02:58:26 +0000 (0:00:08.350)       0:09:26.371 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nThursday 26 March 2026  02:58:31 +0000 (0:00:04.352)       0:09:30.723 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nThursday 26 March 2026  02:58:31 +0000 (0:00:00.718)       0:09:31.442 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nThursday 26 March 2026  02:58:32 +0000 (0:00:01.072)       0:09:32.514 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  02:58:35 +0000 (0:00:02.118)       0:09:34.633 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nThursday 26 March 2026  02:58:36 +0000 (0:00:01.512)       0:09:36.145 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nThursday 26 March 2026  02:58:39 +0000 (0:00:02.526)       0:09:38.672 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nThursday 26 March 2026  02:58:39 +0000 (0:00:00.370)       0:09:39.042 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  02:58:40 +0000 (0:00:00.954)       0:09:39.997 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nThursday 26 March 2026  02:58:41 +0000 (0:00:00.669)       0:09:40.667 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nThursday 26 March 2026  02:58:42 +0000 (0:00:01.277)       0:09:41.944 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nThursday 26 March 2026  02:58:43 +0000 (0:00:01.458)       0:09:43.402 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nThursday 26 March 2026  02:58:45 +0000 (0:00:01.674)       0:09:45.077 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nThursday 26 March 2026  02:58:47 +0000 (0:00:01.827)       0:09:46.904 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  02:58:48 +0000 (0:00:01.559)       0:09:48.464 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  02:58:49 +0000 (0:00:00.669)       0:09:49.133 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nThursday 26 March 2026  02:58:50 +0000 (0:00:00.864)       0:09:49.998 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nThursday 26 March 2026  02:58:50 +0000 (0:00:00.370)       0:09:50.368 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nThursday 26 March 2026  02:58:51 +0000 (0:00:01.114)       0:09:51.482 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nThursday 26 March 2026  02:58:52 +0000 (0:00:00.959)       0:09:52.442 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  02:58:53 +0000 (0:00:00.926)       0:09:53.368 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  02:58:54 +0000 (0:00:00.558)       0:09:53.926 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  02:58:55 +0000 (0:00:00.896)       0:09:54.823 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  02:58:55 +0000 (0:00:00.342)       0:09:55.166 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  02:58:56 +0000 (0:00:00.682)       0:09:55.848 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  02:58:57 +0000 (0:00:01.071)       0:09:56.920 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  02:58:58 +0000 (0:00:00.699)       0:09:57.620 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  02:58:58 +0000 (0:00:00.347)       0:09:57.968 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  02:58:58 +0000 (0:00:00.337)       0:09:58.305 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  02:58:59 +0000 (0:00:00.669)       0:09:58.975 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  02:59:00 +0000 (0:00:00.714)       0:09:59.689 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  02:59:00 +0000 (0:00:00.704)       0:10:00.394 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  02:59:01 +0000 (0:00:00.354)       0:10:00.748 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  02:59:01 +0000 (0:00:00.652)       0:10:01.400 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  02:59:02 +0000 (0:00:00.372)       0:10:01.773 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  02:59:02 +0000 (0:00:00.370)       0:10:02.143 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  02:59:03 +0000 (0:00:00.378)       0:10:02.522 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  02:59:03 +0000 (0:00:00.658)       0:10:03.181 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  02:59:03 +0000 (0:00:00.328)       0:10:03.510 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  02:59:04 +0000 (0:00:00.358)       0:10:03.868 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  02:59:04 +0000 (0:00:00.338)       0:10:04.206 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nThursday 26 March 2026  02:59:05 +0000 (0:00:00.940)       0:10:05.147 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  02:59:06 +0000 (0:00:00.846)       0:10:05.993 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  02:59:08 +0000 (0:00:02.156)       0:10:08.150 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nThursday 26 March 2026  02:59:09 +0000 (0:00:01.227)       0:10:09.378 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nThursday 26 March 2026  02:59:10 +0000 (0:00:00.373)       0:10:09.751 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nThursday 26 March 2026  02:59:11 +0000 (0:00:00.841)       0:10:10.593 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nThursday 26 March 2026  02:59:11 +0000 (0:00:00.815)       0:10:11.408 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  02:59:16 +0000 (0:00:04.560)       0:10:15.969 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  02:59:18 +0000 (0:00:02.437)       0:10:18.407 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nThursday 26 March 2026  02:59:20 +0000 (0:00:01.617)       0:10:20.024 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nThursday 26 March 2026  02:59:20 +0000 (0:00:00.267)       0:10:20.292 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nThursday 26 March 2026  02:59:21 +0000 (0:00:00.660)       0:10:20.953 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nThursday 26 March 2026  02:59:22 +0000 (0:00:00.660)       0:10:21.614 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nThursday 26 March 2026  02:59:53 +0000 (0:00:31.288)       0:10:52.902 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nThursday 26 March 2026  02:59:53 +0000 (0:00:00.365)       0:10:53.268 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nThursday 26 March 2026  02:59:54 +0000 (0:00:00.610)       0:10:53.878 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nThursday 26 March 2026  02:59:55 +0000 (0:00:00.648)       0:10:54.526 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nThursday 26 March 2026  02:59:55 +0000 (0:00:00.915)       0:10:55.442 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nThursday 26 March 2026  02:59:57 +0000 (0:00:01.331)       0:10:56.773 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nThursday 26 March 2026  02:59:58 +0000 (0:00:01.234)       0:10:58.008 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nThursday 26 March 2026  03:00:00 +0000 (0:00:01.817)       0:10:59.825 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  03:00:03 +0000 (0:00:02.837)       0:11:02.663 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  03:00:03 +0000 (0:00:00.373)       0:11:03.036 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nThursday 26 March 2026  03:00:04 +0000 (0:00:00.907)       0:11:03.943 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nThursday 26 March 2026  03:00:04 +0000 (0:00:00.422)       0:11:04.365 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nThursday 26 March 2026  03:00:05 +0000 (0:00:00.521)       0:11:04.887 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nThursday 26 March 2026  03:00:06 +0000 (0:00:01.352)       0:11:06.239 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:00:06 +0000 (0:00:00.276)       0:11:06.515 ******** \n=============================================================================== \nceph-container-common : Pulling Ceph container image ------------------- 64.17s\nceph-osd : Use ceph-volume to create osds ------------------------------ 41.01s\nceph-rgw : Create rgw pools -------------------------------------------- 31.29s\nceph-mgr : Wait for all mgr to be up ----------------------------------- 24.28s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.85s\nceph-mon : Set cluster configs ----------------------------------------- 14.92s\nceph-osd : Wait for all osd to be up ----------------------------------- 12.70s\nceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.63s\nceph-mon : Fetch ceph initial keys -------------------------------------- 8.99s\nceph-mds : Create filesystem pools -------------------------------------- 8.35s\nceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.37s\nceph-config : Create ceph initial directories --------------------------- 6.36s\nceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.32s\nceph-rgw : Create rgw keyrings ------------------------------------------ 4.56s\nceph-mds : Create ceph filesystem --------------------------------------- 4.35s\nceph-crash : Create client.crash keyring -------------------------------- 4.20s\nceph-crash : Start the ceph-crash service ------------------------------- 3.72s\nceph-container-common : Get ceph version -------------------------------- 3.47s\nceph-osd : Systemd start osd -------------------------------------------- 3.43s\nceph-mon : Copy admin keyring over to mons ------------------------------ 3.13s\n\u001b[32m2026-03-26 03:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62df255a-8406-4b80-a8bf-efaf3055d2ba (ceph-pools) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62df255a-8406-4b80-a8bf-efaf3055d2ba (ceph-pools) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Create ceph pools] *******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  03:00:15 +0000 (0:00:00.682)       0:00:00.682 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  03:00:16 +0000 (0:00:00.729)       0:00:01.411 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  03:00:16 +0000 (0:00:00.672)       0:00:02.083 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  03:00:17 +0000 (0:00:00.346)       0:00:02.430 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  03:00:18 +0000 (0:00:00.953)       0:00:03.384 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  03:00:18 +0000 (0:00:00.372)       0:00:03.756 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  03:00:18 +0000 (0:00:00.309)       0:00:04.066 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  03:00:19 +0000 (0:00:00.344)       0:00:04.410 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  03:00:19 +0000 (0:00:00.607)       0:00:05.018 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  03:00:19 +0000 (0:00:00.320)       0:00:05.339 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  03:00:20 +0000 (0:00:00.675)       0:00:06.014 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  03:00:21 +0000 (0:00:00.504)       0:00:06.518 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  03:00:23 +0000 (0:00:02.274)       0:00:08.793 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  03:00:24 +0000 (0:00:00.695)       0:00:09.488 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  03:00:25 +0000 (0:00:01.166)       0:00:10.654 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  03:00:25 +0000 (0:00:00.178)       0:00:10.833 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 03:00:22.131422', 'end': '2026-03-26 03:00:22.186955', 'delta': '0:00:00.055533', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 03:00:22.709658', 'end': '2026-03-26 03:00:22.757960', 'delta': '0:00:00.048302', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 03:00:23.269066', 'end': '2026-03-26 03:00:23.309660', 'delta': '0:00:00.040594', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  03:00:25 +0000 (0:00:00.199)       0:00:11.032 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  03:00:26 +0000 (0:00:00.455)       0:00:11.488 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  03:00:27 +0000 (0:00:01.685)       0:00:13.174 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  03:00:28 +0000 (0:00:00.362)       0:00:13.536 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  03:00:28 +0000 (0:00:00.741)       0:00:14.277 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  03:00:29 +0000 (0:00:00.345)       0:00:14.623 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  03:00:29 +0000 (0:00:00.162)       0:00:14.785 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  03:00:29 +0000 (0:00:00.263)       0:00:15.048 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  03:00:30 +0000 (0:00:00.378)       0:00:15.427 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  03:00:30 +0000 (0:00:00.601)       0:00:16.028 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  03:00:31 +0000 (0:00:00.353)       0:00:16.381 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  03:00:31 +0000 (0:00:00.370)       0:00:16.752 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  03:00:31 +0000 (0:00:00.344)       0:00:17.096 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  03:00:32 +0000 (0:00:00.537)       0:00:17.633 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  03:00:32 +0000 (0:00:00.358)       0:00:17.992 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  03:00:33 +0000 (0:00:00.693)       0:00:18.685 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  03:00:33 +0000 (0:00:00.638)       0:00:19.323 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  03:00:34 +0000 (0:00:01.004)       0:00:20.328 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  03:00:35 +0000 (0:00:00.323)       0:00:20.652 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  03:00:35 +0000 (0:00:00.673)       0:00:21.326 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  03:00:36 +0000 (0:00:00.312)       0:00:21.638 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  03:00:37 +0000 (0:00:00.703)       0:00:22.342 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  03:00:37 +0000 (0:00:00.372)       0:00:22.715 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  03:00:38 +0000 (0:00:01.134)       0:00:23.849 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  03:00:38 +0000 (0:00:00.419)       0:00:24.269 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  03:00:39 +0000 (0:00:00.802)       0:00:25.072 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  03:00:40 +0000 (0:00:00.331)       0:00:25.404 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  03:00:40 +0000 (0:00:00.329)       0:00:25.734 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  03:00:40 +0000 (0:00:00.595)       0:00:26.329 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  03:00:41 +0000 (0:00:00.440)       0:00:26.769 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  03:00:41 +0000 (0:00:00.407)       0:00:27.176 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  03:00:42 +0000 (0:00:00.375)       0:00:27.552 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  03:00:42 +0000 (0:00:00.426)       0:00:27.978 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  03:00:42 +0000 (0:00:00.347)       0:00:28.326 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  03:00:43 +0000 (0:00:00.839)       0:00:29.166 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  03:00:44 +0000 (0:00:00.872)       0:00:30.038 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Include tasks from the ceph-osd role] ************************************\nThursday 26 March 2026  03:00:46 +0000 (0:00:01.762)       0:00:31.801 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m\n\nTASK [create openstack pool(s)] ************************************************\nThursday 26 March 2026  03:00:47 +0000 (0:00:00.617)       0:00:32.418 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [generate keys] ***********************************************************\nThursday 26 March 2026  03:01:33 +0000 (0:00:46.350)       0:01:18.768 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [get keys from monitors] **************************************************\nThursday 26 March 2026  03:01:57 +0000 (0:00:23.918)       0:01:42.687 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [copy ceph key(s) if needed] **********************************************\nThursday 26 March 2026  03:02:08 +0000 (0:00:11.575)       0:01:54.262 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:02:26 +0000 (0:00:17.220)       0:02:11.483 ******** \n=============================================================================== \ncreate openstack pool(s) ----------------------------------------------- 46.35s\ngenerate keys ---------------------------------------------------------- 23.92s\ncopy ceph key(s) if needed --------------------------------------------- 17.22s\nget keys from monitors ------------------------------------------------- 11.58s\nceph-facts : Find a running mon container ------------------------------- 2.27s\nceph-facts : Set_fact ceph_admin_command -------------------------------- 1.76s\nceph-facts : Get current fsid if cluster is already running ------------- 1.69s\nceph-facts : Check if the ceph mon socket is in-use --------------------- 1.17s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 1.13s\nceph-facts : Check if the ceph conf exists ------------------------------ 1.00s\nceph-facts : Check if podman binary is present -------------------------- 0.95s\nceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.87s\nceph-facts : Set_fact rgw_instances ------------------------------------- 0.84s\nceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.80s\nceph-facts : Get current fsid ------------------------------------------- 0.74s\nceph-facts : Include facts.yml ------------------------------------------ 0.73s\nceph-facts : Read osd pool default crush rule --------------------------- 0.70s\nceph-facts : Check for a ceph mon socket -------------------------------- 0.70s\nceph-facts : Collect existed devices ------------------------------------ 0.69s\nceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.68s\n\u001b[32m2026-03-26 03:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba06c60f-34f0-4428-a23d-714319bfb192 (copy-ceph-keys) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ba06c60f-34f0-4428-a23d-714319bfb192 (copy-ceph-keys) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy ceph keys to the configuration repository] **************************\n\nTASK [Check if ceph keys exist] ************************************************\nThursday 26 March 2026  03:02:33 +0000 (0:00:00.164)       0:00:00.164 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Fetch all ceph keys] *****************************************************\nThursday 26 March 2026  03:02:38 +0000 (0:00:04.691)       0:00:04.855 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Create share directory] **************************************************\nThursday 26 March 2026  03:02:42 +0000 (0:00:04.424)       0:00:09.280 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Write ceph keys to the share directory] **********************************\nThursday 26 March 2026  03:02:44 +0000 (0:00:01.072)       0:00:10.353 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Check if target directories exist] ***************************************\nThursday 26 March 2026  03:02:58 +0000 (0:00:14.271)       0:00:24.624 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m\n\nTASK [Write ceph keys to the configuration directory] **************************\nThursday 26 March 2026  03:03:02 +0000 (0:00:04.262)       0:00:28.887 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:03:09 +0000 (0:00:07.279)       0:00:36.166 ******** \n=============================================================================== \nWrite ceph keys to the share directory --------------------------------- 14.27s\nWrite ceph keys to the configuration directory -------------------------- 7.28s\nCheck if ceph keys exist ------------------------------------------------ 4.69s\nFetch all ceph keys ----------------------------------------------------- 4.42s\nCheck if target directories exist --------------------------------------- 4.26s\nCreate share directory -------------------------------------------------- 1.07s\n\u001b[32m2026-03-26 03:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3617dc68-3c70-4fe1-b924-b0629be23217 (cephclient) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3617dc68-3c70-4fe1-b924-b0629be23217 (cephclient) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nThursday 26 March 2026  03:03:27 +0000 (0:00:00.256)       0:00:00.256 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nThursday 26 March 2026  03:03:27 +0000 (0:00:00.265)       0:00:00.522 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nThursday 26 March 2026  03:03:28 +0000 (0:00:01.247)       0:00:01.769 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nThursday 26 March 2026  03:03:30 +0000 (0:00:01.495)       0:00:03.265 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nThursday 26 March 2026  03:03:31 +0000 (0:00:00.984)       0:00:04.249 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nThursday 26 March 2026  03:03:32 +0000 (0:00:00.973)       0:00:05.223 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nThursday 26 March 2026  03:04:15 +0000 (0:00:43.450)       0:00:48.673 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nThursday 26 March 2026  03:04:19 +0000 (0:00:04.299)       0:00:52.973 ******** \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nThursday 26 March 2026  03:04:20 +0000 (0:00:00.500)       0:00:53.473 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nThursday 26 March 2026  03:04:20 +0000 (0:00:00.158)       0:00:53.631 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******\nThursday 26 March 2026  03:04:21 +0000 (0:00:00.574)       0:00:54.205 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***\nThursday 26 March 2026  03:04:22 +0000 (0:00:01.680)       0:00:55.885 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******\nThursday 26 March 2026  03:04:23 +0000 (0:00:00.705)       0:00:56.591 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****\nThursday 26 March 2026  03:04:24 +0000 (0:00:00.622)       0:00:57.214 ******** \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:04:25 +0000 (0:00:01.594)       0:00:58.809 ******** \n=============================================================================== \nosism.services.cephclient : Manage cephclient service ------------------ 43.45s\nosism.services.cephclient : Copy wrapper scripts ------------------------ 4.30s\nosism.services.cephclient : Restart cephclient service ------------------ 1.68s\nosism.services.cephclient : Copy bash completion scripts ---------------- 1.59s\nosism.services.cephclient : Copy configuration files -------------------- 1.50s\nosism.services.cephclient : Create required directories ----------------- 1.25s\nosism.services.cephclient : Copy keyring file --------------------------- 0.98s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 0.97s\nosism.services.cephclient : Ensure that all containers are up ----------- 0.71s\nosism.services.cephclient : Wait for an healthy service ----------------- 0.62s\nosism.services.cephclient : Include rook task --------------------------- 0.57s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.50s\nosism.services.cephclient : Include container tasks --------------------- 0.27s\nosism.services.cephclient : Include package tasks ----------------------- 0.16s\n\u001b[32m2026-03-26 03:04:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccda482a-8362-4afe-bbda-d062fa374a95 (ceph-bootstrap-dashboard) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:04:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ccda482a-8362-4afe-bbda-d062fa374a95 (ceph-bootstrap-dashboard) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Bootstraph ceph dashboard] ***********************************************\n\nTASK [Disable the ceph dashboard] **********************************************\nThursday 26 March 2026  03:04:33 +0000 (0:00:00.273)       0:00:00.273 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/ssl to false] ******************************************\nThursday 26 March 2026  03:04:35 +0000 (0:00:01.841)       0:00:02.115 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_port to 7000] ***********************************\nThursday 26 March 2026  03:04:36 +0000 (0:00:01.161)       0:00:03.276 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************\nThursday 26 March 2026  03:04:37 +0000 (0:00:01.144)       0:00:04.420 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_behaviour to error] ****************************\nThursday 26 March 2026  03:04:38 +0000 (0:00:01.271)       0:00:05.692 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_error_status_code to 404] **********************\nThursday 26 March 2026  03:04:40 +0000 (0:00:01.123)       0:00:06.816 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable the ceph dashboard] ***********************************************\nThursday 26 March 2026  03:04:41 +0000 (0:00:01.124)       0:00:07.941 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Write ceph_dashboard_password to temporary file] *************************\nThursday 26 March 2026  03:04:43 +0000 (0:00:02.107)       0:00:10.048 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Create admin user] *******************************************************\nThursday 26 March 2026  03:04:44 +0000 (0:00:01.337)       0:00:11.386 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Remove temporary file for ceph_dashboard_password] ***********************\nThursday 26 March 2026  03:05:25 +0000 (0:00:40.591)       0:00:51.978 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nThursday 26 March 2026  03:05:25 +0000 (0:00:00.170)       0:00:52.149 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nThursday 26 March 2026  03:05:37 +0000 (0:00:12.040)       0:01:04.189 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nThursday 26 March 2026  03:05:38 +0000 (0:00:01.252)       0:01:05.441 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:05:50 +0000 (0:00:11.365)       0:01:16.807 ******** \n=============================================================================== \nCreate admin user ------------------------------------------------------ 40.59s\nRestart ceph manager service ------------------------------------------- 24.66s\nEnable the ceph dashboard ----------------------------------------------- 2.11s\nDisable the ceph dashboard ---------------------------------------------- 1.84s\nWrite ceph_dashboard_password to temporary file ------------------------- 1.34s\nSet mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.27s\nSet mgr/dashboard/ssl to false ------------------------------------------ 1.16s\nSet mgr/dashboard/server_port to 7000 ----------------------------------- 1.14s\nSet mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.12s\nSet mgr/dashboard/standby_behaviour to error ---------------------------- 1.12s\nRemove temporary file for ceph_dashboard_password ----------------------- 0.17s\n+ sh -c /opt/configuration/scripts/deploy/300-openstack.sh\n\u001b[32m2026-03-26 03:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5d3630be-9a4e-48be-9e20-a75775b77781 (keystone) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5d3630be-9a4e-48be-9e20-a75775b77781 (keystone) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:05:57 +0000 (0:00:00.282)       0:00:00.282 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:05:57 +0000 (0:00:00.343)       0:00:00.625 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Apply role keystone] *****************************************************\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:05:58 +0000 (0:00:00.442)       0:00:01.067 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Ensuring config directories exist] ****************************\nThursday 26 March 2026  03:05:58 +0000 (0:00:00.605)       0:00:01.673 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Check if policies shall be overwritten] ***********************\nThursday 26 March 2026  03:06:00 +0000 (0:00:01.621)       0:00:03.294 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Set keystone policy file] *************************************\nThursday 26 March 2026  03:06:00 +0000 (0:00:00.323)       0:00:03.618 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Check if Keystone domain-specific config is supplied] *********\nThursday 26 March 2026  03:06:01 +0000 (0:00:00.342)       0:00:03.961 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:06:01 +0000 (0:00:00.874)       0:00:04.835 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over extra CA certificates] *******\nThursday 26 March 2026  03:06:02 +0000 (0:00:00.623)       0:00:05.459 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:06:06 +0000 (0:00:03.477)       0:00:08.936 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****\nThursday 26 March 2026  03:06:06 +0000 (0:00:00.583)       0:00:09.520 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over config.json files for services] ******************\nThursday 26 March 2026  03:06:07 +0000 (0:00:00.888)       0:00:10.408 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying over keystone.conf] ***********************************\nThursday 26 March 2026  03:06:10 +0000 (0:00:03.322)       0:00:13.730 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying keystone-startup script for keystone] *****************\nThursday 26 March 2026  03:06:15 +0000 (0:00:04.719)       0:00:18.450 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Create Keystone domain-specific config directory] *************\nThursday 26 March 2026  03:06:16 +0000 (0:00:01.429)       0:00:19.880 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Get file list in custom domains folder] ***********************\nThursday 26 March 2026  03:06:17 +0000 (0:00:00.806)       0:00:20.686 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying Keystone Domain specific settings] ********************\nThursday 26 March 2026  03:06:18 +0000 (0:00:00.569)       0:00:21.256 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over existing policy file] ****************************\nThursday 26 March 2026  03:06:18 +0000 (0:00:00.323)       0:00:21.579 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:06:19 +0000 (0:00:00.588)       0:00:22.168 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over wsgi-keystone.conf] ******************************\nThursday 26 March 2026  03:06:19 +0000 (0:00:00.309)       0:00:22.477 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\nTASK [keystone : Checking whether keystone-paste.ini file exists] **************\nThursday 26 March 2026  03:06:21 +0000 (0:00:01.859)       0:00:24.336 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Copying over keystone-paste.ini] ******************************\nThursday 26 March 2026  03:06:22 +0000 (0:00:00.953)       0:00:25.290 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Generate the required cron jobs for the node] *****************\nThursday 26 March 2026  03:06:22 +0000 (0:00:00.601)       0:00:25.891 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\nTASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***\nThursday 26 March 2026  03:06:24 +0000 (0:00:01.090)       0:00:26.982 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying files for keystone-fernet] ****************************\nThursday 26 March 2026  03:06:24 +0000 (0:00:00.565)       0:00:27.547 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [keystone : Copying files for keystone-ssh] *******************************\nThursday 26 March 2026  03:06:33 +0000 (0:00:09.154)       0:00:36.701 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\nTASK [keystone : Check keystone containers] ************************************\nThursday 26 March 2026  03:06:36 +0000 (0:00:02.672)       0:00:39.373 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:06:38 +0000 (0:00:02.245)       0:00:41.619 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Creating keystone database] ***********************************\nThursday 26 March 2026  03:06:39 +0000 (0:00:00.588)       0:00:42.207 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating Keystone database user and setting permissions] ******\nThursday 26 March 2026  03:06:41 +0000 (0:00:02.362)       0:00:44.570 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Checking for any running keystone_fernet containers] **********\nThursday 26 March 2026  03:06:43 +0000 (0:00:02.231)       0:00:46.802 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Group nodes where keystone_fernet is running] *****************\nThursday 26 March 2026  03:06:44 +0000 (0:00:00.871)       0:00:47.673 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***\nThursday 26 March 2026  03:06:45 +0000 (0:00:00.388)       0:00:48.062 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Running Keystone bootstrap container] *************************\nThursday 26 March 2026  03:06:45 +0000 (0:00:00.582)       0:00:48.644 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Running Keystone fernet bootstrap container] ******************\nThursday 26 March 2026  03:07:00 +0000 (0:00:15.232)       0:01:03.876 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Flush handlers] ***********************************************\nThursday 26 March 2026  03:07:12 +0000 (0:00:11.133)       0:01:15.010 ******** \n\nTASK [keystone : Flush handlers] ***********************************************\nThursday 26 March 2026  03:07:12 +0000 (0:00:00.085)       0:01:15.096 ******** \n\nTASK [keystone : Flush handlers] ***********************************************\nThursday 26 March 2026  03:07:12 +0000 (0:00:00.072)       0:01:15.169 ******** \n\nRUNNING HANDLER [keystone : Restart keystone-ssh container] ********************\nThursday 26 March 2026  03:07:12 +0000 (0:00:00.072)       0:01:15.241 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone-fernet container] *****************\nThursday 26 March 2026  03:08:01 +0000 (0:00:49.106)       0:02:04.348 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone container] ************************\nThursday 26 March 2026  03:08:11 +0000 (0:00:10.443)       0:02:14.791 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:08:19 +0000 (0:00:07.404)       0:02:22.195 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Waiting for Keystone SSH port to be UP] ***********************\nThursday 26 March 2026  03:08:19 +0000 (0:00:00.644)       0:02:22.840 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Run key distribution] *****************************************\nThursday 26 March 2026  03:08:21 +0000 (0:00:01.178)       0:02:24.018 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating admin project, user, role, service, and endpoint] ****\nThursday 26 March 2026  03:08:22 +0000 (0:00:01.778)       0:02:25.797 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating services] **********************\nThursday 26 March 2026  03:08:34 +0000 (0:00:11.928)       0:02:37.725 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m\n\nTASK [service-ks-register : keystone | Creating endpoints] *********************\nThursday 26 March 2026  03:08:59 +0000 (0:00:24.419)       0:03:02.145 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating projects] **********************\nThursday 26 March 2026  03:09:06 +0000 (0:00:06.757)       0:03:08.902 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating users] *************************\nThursday 26 March 2026  03:09:06 +0000 (0:00:00.134)       0:03:09.036 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating roles] *************************\nThursday 26 March 2026  03:09:06 +0000 (0:00:00.137)       0:03:09.174 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Granting user roles] ********************\nThursday 26 March 2026  03:09:06 +0000 (0:00:00.139)       0:03:09.313 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating default user role] ***********************************\nThursday 26 March 2026  03:09:06 +0000 (0:00:00.557)       0:03:09.871 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nThursday 26 March 2026  03:09:10 +0000 (0:00:03.281)       0:03:13.153 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:09:10 +0000 (0:00:00.512)       0:03:13.665 ******** \n=============================================================================== \nkeystone : Restart keystone-ssh container ------------------------------ 49.11s\nservice-ks-register : keystone | Creating services --------------------- 24.42s\nkeystone : Running Keystone bootstrap container ------------------------ 15.23s\nkeystone : Creating admin project, user, role, service, and endpoint --- 11.93s\nkeystone : Running Keystone fernet bootstrap container ----------------- 11.13s\nkeystone : Restart keystone-fernet container --------------------------- 10.44s\nkeystone : Copying files for keystone-fernet ---------------------------- 9.15s\nkeystone : Restart keystone container ----------------------------------- 7.40s\nservice-ks-register : keystone | Creating endpoints --------------------- 6.76s\nkeystone : Copying over keystone.conf ----------------------------------- 4.72s\nservice-cert-copy : keystone | Copying over extra CA certificates ------- 3.48s\nkeystone : Copying over config.json files for services ------------------ 3.32s\nkeystone : Creating default user role ----------------------------------- 3.28s\nkeystone : Copying files for keystone-ssh ------------------------------- 2.67s\nkeystone : Creating keystone database ----------------------------------- 2.36s\nkeystone : Check keystone containers ------------------------------------ 2.25s\nkeystone : Creating Keystone database user and setting permissions ------ 2.23s\nkeystone : Copying over wsgi-keystone.conf ------------------------------ 1.86s\nkeystone : Run key distribution ----------------------------------------- 1.78s\nkeystone : Ensuring config directories exist ---------------------------- 1.62s\n\u001b[32m2026-03-26 03:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fb909599-55bc-4ffe-8cce-71be6ebf1872 (placement) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fb909599-55bc-4ffe-8cce-71be6ebf1872 (placement) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:09:18 +0000 (0:00:00.270)       0:00:00.270 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:09:18 +0000 (0:00:00.316)       0:00:00.587 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m\n\nPLAY [Apply role placement] ****************************************************\n\nTASK [placement : include_tasks] ***********************************************\nThursday 26 March 2026  03:09:19 +0000 (0:00:00.474)       0:00:01.061 ******** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : placement | Creating services] *********************\nThursday 26 March 2026  03:09:19 +0000 (0:00:00.581)       0:00:01.642 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m\n\nTASK [service-ks-register : placement | Creating endpoints] ********************\nThursday 26 March 2026  03:09:23 +0000 (0:00:03.627)       0:00:05.269 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m\n\nTASK [service-ks-register : placement | Creating projects] *********************\nThursday 26 March 2026  03:09:29 +0000 (0:00:06.439)       0:00:11.709 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating users] ************************\nThursday 26 March 2026  03:09:33 +0000 (0:00:03.547)       0:00:15.256 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating roles] ************************\nThursday 26 March 2026  03:09:37 +0000 (0:00:04.251)       0:00:19.507 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : placement | Granting user roles] *******************\nThursday 26 March 2026  03:09:40 +0000 (0:00:03.104)       0:00:22.612 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nThursday 26 March 2026  03:09:44 +0000 (0:00:04.224)       0:00:26.837 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Ensuring config directories exist] ***************************\nThursday 26 March 2026  03:09:45 +0000 (0:00:00.312)       0:00:27.150 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Check if policies shall be overwritten] **********************\nThursday 26 March 2026  03:09:46 +0000 (0:00:01.155)       0:00:28.305 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [placement : Set placement policy file] ***********************************\nThursday 26 March 2026  03:09:46 +0000 (0:00:00.371)       0:00:28.677 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nThursday 26 March 2026  03:09:47 +0000 (0:00:00.350)       0:00:29.028 ******** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over extra CA certificates] ******\nThursday 26 March 2026  03:09:47 +0000 (0:00:00.583)       0:00:29.611 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:09:49 +0000 (0:00:01.759)       0:00:31.371 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS key] ***\nThursday 26 March 2026  03:09:50 +0000 (0:00:00.604)       0:00:31.975 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over config.json files for services] *****************\nThursday 26 March 2026  03:09:50 +0000 (0:00:00.753)       0:00:32.729 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement.conf] *********************************\nThursday 26 March 2026  03:09:52 +0000 (0:00:01.693)       0:00:34.423 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement-api wsgi configuration] ***************\nThursday 26 March 2026  03:09:54 +0000 (0:00:02.350)       0:00:36.773 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\nTASK [placement : Copying over migrate-db.rc.j2 configuration] *****************\nThursday 26 March 2026  03:09:56 +0000 (0:00:01.519)       0:00:38.293 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over existing policy file] ***************************\nThursday 26 March 2026  03:09:57 +0000 (0:00:01.311)       0:00:39.605 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Check placement containers] **********************************\nThursday 26 March 2026  03:09:58 +0000 (0:00:00.803)       0:00:40.408 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Creating placement databases] ********************************\nThursday 26 March 2026  03:09:59 +0000 (0:00:01.119)       0:00:41.528 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Creating placement databases user and setting permissions] ***\nThursday 26 March 2026  03:10:01 +0000 (0:00:02.081)       0:00:43.609 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Running placement bootstrap container] ***********************\nThursday 26 March 2026  03:10:04 +0000 (0:00:02.366)       0:00:45.976 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Flush handlers] **********************************************\nThursday 26 March 2026  03:10:18 +0000 (0:00:14.182)       0:01:00.158 ******** \n\nTASK [placement : Flush handlers] **********************************************\nThursday 26 March 2026  03:10:18 +0000 (0:00:00.076)       0:01:00.234 ******** \n\nTASK [placement : Flush handlers] **********************************************\nThursday 26 March 2026  03:10:18 +0000 (0:00:00.074)       0:01:00.308 ******** \n\nRUNNING HANDLER [placement : Restart placement-api container] ******************\nThursday 26 March 2026  03:10:18 +0000 (0:00:00.071)       0:01:00.380 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:10:28 +0000 (0:00:10.449)       0:01:10.829 ******** \n=============================================================================== \nplacement : Running placement bootstrap container ---------------------- 14.18s\nplacement : Restart placement-api container ---------------------------- 10.45s\nservice-ks-register : placement | Creating endpoints -------------------- 6.44s\nservice-ks-register : placement | Creating users ------------------------ 4.25s\nservice-ks-register : placement | Granting user roles ------------------- 4.22s\nservice-ks-register : placement | Creating services --------------------- 3.63s\nservice-ks-register : placement | Creating projects --------------------- 3.55s\nservice-ks-register : placement | Creating roles ------------------------ 3.10s\nplacement : Creating placement databases user and setting permissions --- 2.37s\nplacement : Copying over placement.conf --------------------------------- 2.35s\nplacement : Creating placement databases -------------------------------- 2.08s\nservice-cert-copy : placement | Copying over extra CA certificates ------ 1.76s\nplacement : Copying over config.json files for services ----------------- 1.69s\nplacement : Copying over placement-api wsgi configuration --------------- 1.52s\nplacement : Copying over migrate-db.rc.j2 configuration ----------------- 1.31s\nplacement : Ensuring config directories exist --------------------------- 1.16s\nplacement : Check placement containers ---------------------------------- 1.12s\nplacement : Copying over existing policy file --------------------------- 0.80s\nservice-cert-copy : placement | Copying over backend internal TLS key --- 0.75s\nservice-cert-copy : placement | Copying over backend internal TLS certificate --- 0.60s\n\u001b[32m2026-03-26 03:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0cd80ac4-a396-4ee1-bd88-7ddd0a9d3192 (neutron) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0cd80ac4-a396-4ee1-bd88-7ddd0a9d3192 (neutron) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:10:36 +0000 (0:00:00.296)       0:00:00.296 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:10:37 +0000 (0:00:00.730)       0:00:01.027 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m\n\nPLAY [Apply role neutron] ******************************************************\n\nTASK [neutron : include_tasks] *************************************************\nThursday 26 March 2026  03:10:37 +0000 (0:00:00.670)       0:00:01.697 ******** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [neutron : Get container facts] *******************************************\nThursday 26 March 2026  03:10:39 +0000 (0:00:01.487)       0:00:03.185 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Get container volume facts] ************************************\nThursday 26 March 2026  03:10:40 +0000 (0:00:01.369)       0:00:04.554 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check for ML2/OVN presence] ************************************\nThursday 26 March 2026  03:10:41 +0000 (0:00:01.138)       0:00:05.692 ******** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [neutron : Check for ML2/OVS presence] ************************************\nThursday 26 March 2026  03:10:42 +0000 (0:00:00.915)       0:00:06.608 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-ks-register : neutron | Creating services] ***********************\nThursday 26 March 2026  03:10:43 +0000 (0:00:00.645)       0:00:07.254 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m\n\nTASK [service-ks-register : neutron | Creating endpoints] **********************\nThursday 26 March 2026  03:10:47 +0000 (0:00:03.757)       0:00:11.012 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating projects] ***********************\nThursday 26 March 2026  03:10:53 +0000 (0:00:06.386)       0:00:17.399 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating users] **************************\nThursday 26 March 2026  03:10:56 +0000 (0:00:03.132)       0:00:20.531 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating roles] **************************\nThursday 26 March 2026  03:11:00 +0000 (0:00:03.842)       0:00:24.373 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : neutron | Granting user roles] *********************\nThursday 26 March 2026  03:11:03 +0000 (0:00:03.090)       0:00:27.464 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nThursday 26 March 2026  03:11:11 +0000 (0:00:08.310)       0:00:35.774 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Load and persist kernel modules] *****************************************\nThursday 26 March 2026  03:11:12 +0000 (0:00:00.839)       0:00:36.613 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check IPv6 support] ********************************************\nThursday 26 March 2026  03:11:14 +0000 (0:00:02.199)       0:00:38.813 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nThursday 26 March 2026  03:11:16 +0000 (0:00:01.211)       0:00:40.024 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Ensuring config directories exist] *****************************\nThursday 26 March 2026  03:11:18 +0000 (0:00:02.297)       0:00:42.322 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Check if extra ml2 plugins exists] *****************************\nThursday 26 March 2026  03:11:21 +0000 (0:00:02.642)       0:00:44.965 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m\n\u001b[1;35ma directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nThursday 26 March 2026  03:11:21 +0000 (0:00:00.854)       0:00:45.819 ******** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over extra CA certificates] ********\nThursday 26 March 2026  03:11:23 +0000 (0:00:01.359)       0:00:47.179 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:11:26 +0000 (0:00:03.511)       0:00:50.690 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****\nThursday 26 March 2026  03:11:28 +0000 (0:00:02.225)       0:00:52.916 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating TLS backend PEM File] *********************************\nThursday 26 March 2026  03:11:32 +0000 (0:00:03.054)       0:00:55.971 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:11:34 +0000 (0:00:02.514)       0:00:58.485 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Set neutron policy file] ***************************************\nThursday 26 March 2026  03:11:34 +0000 (0:00:00.171)       0:00:58.656 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over existing policy file] *****************************\nThursday 26 March 2026  03:11:35 +0000 (0:00:00.663)       0:00:59.319 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over config.json files for services] *******************\nThursday 26 March 2026  03:11:37 +0000 (0:00:02.568)       0:01:01.888 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron.conf] *************************************\nThursday 26 March 2026  03:11:41 +0000 (0:00:03.393)       0:01:05.281 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron_vpnaas.conf] ******************************\nThursday 26 March 2026  03:11:47 +0000 (0:00:06.438)       0:01:11.719 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ssh key] ******************************************\nThursday 26 March 2026  03:11:49 +0000 (0:00:02.044)       0:01:13.764 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [neutron : Copying over ml2_conf.ini] *************************************\nThursday 26 March 2026  03:11:52 +0000 (0:00:02.729)       0:01:16.493 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\nTASK [neutron : Copying over linuxbridge_agent.ini] ****************************\nThursday 26 March 2026  03:11:56 +0000 (0:00:03.582)       0:01:20.076 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over openvswitch_agent.ini] ****************************\nThursday 26 March 2026  03:11:58 +0000 (0:00:02.306)       0:01:22.383 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over sriov_agent.ini] **********************************\nThursday 26 March 2026  03:12:00 +0000 (0:00:02.333)       0:01:24.717 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over mlnx_agent.ini] ***********************************\nThursday 26 March 2026  03:12:03 +0000 (0:00:02.410)       0:01:27.127 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over eswitchd.conf] ************************************\nThursday 26 March 2026  03:12:05 +0000 (0:00:02.633)       0:01:29.760 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over dhcp_agent.ini] ***********************************\nThursday 26 March 2026  03:12:08 +0000 (0:00:02.619)       0:01:32.380 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over dnsmasq.conf] *************************************\nThursday 26 March 2026  03:12:10 +0000 (0:00:02.367)       0:01:34.748 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over l3_agent.ini] *************************************\nThursday 26 March 2026  03:12:13 +0000 (0:00:02.958)       0:01:37.706 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over fwaas_driver.ini] *********************************\nThursday 26 March 2026  03:12:16 +0000 (0:00:02.551)       0:01:40.257 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metadata_agent.ini] *******************************\nThursday 26 March 2026  03:12:18 +0000 (0:00:02.376)       0:01:42.633 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************\nThursday 26 March 2026  03:12:20 +0000 (0:00:02.208)       0:01:44.842 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metering_agent.ini] *******************************\nThursday 26 March 2026  03:12:24 +0000 (0:00:03.895)       0:01:48.737 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over ironic_neutron_agent.ini] *************************\nThursday 26 March 2026  03:12:27 +0000 (0:00:02.286)       0:01:51.024 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over bgp_dragent.ini] **********************************\nThursday 26 March 2026  03:12:29 +0000 (0:00:02.426)       0:01:53.450 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ovn_agent.ini] ************************************\nThursday 26 March 2026  03:12:31 +0000 (0:00:02.371)       0:01:55.822 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over nsx.ini] ******************************************\nThursday 26 March 2026  03:12:34 +0000 (0:00:02.504)       0:01:58.327 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copy neutron-l3-agent-wrapper script] **************************\nThursday 26 March 2026  03:12:36 +0000 (0:00:02.256)       0:02:00.583 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over extra ml2 plugins] ********************************\nThursday 26 March 2026  03:12:39 +0000 (0:00:02.555)       0:02:03.138 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************\nThursday 26 March 2026  03:12:41 +0000 (0:00:02.440)       0:02:05.579 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over neutron_taas.conf] ********************************\nThursday 26 March 2026  03:12:43 +0000 (0:00:01.931)       0:02:07.510 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check neutron containers] **************************************\nThursday 26 March 2026  03:12:46 +0000 (0:00:02.528)       0:02:10.039 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nThursday 26 March 2026  03:12:48 +0000 (0:00:02.731)       0:02:12.770 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating Neutron database] *************************************\nThursday 26 March 2026  03:12:49 +0000 (0:00:00.822)       0:02:13.593 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Creating Neutron database user and setting permissions] ********\nThursday 26 March 2026  03:12:51 +0000 (0:00:02.097)       0:02:15.690 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Running Neutron bootstrap container] ***************************\nThursday 26 March 2026  03:12:54 +0000 (0:00:02.247)       0:02:17.938 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:43.340)       0:03:01.278 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.093)       0:03:01.372 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.073)       0:03:01.446 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.074)       0:03:01.520 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.069)       0:03:01.590 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.070)       0:03:01.661 ******** \n\nRUNNING HANDLER [neutron : Restart neutron-server container] *******************\nThursday 26 March 2026  03:13:37 +0000 (0:00:00.072)       0:03:01.733 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******\nThursday 26 March 2026  03:14:07 +0000 (0:00:30.095)       0:03:31.829 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:15:10 +0000 (0:01:02.526)       0:04:34.356 ******** \n=============================================================================== \nneutron : Restart neutron-ovn-metadata-agent container ----------------- 62.53s\nneutron : Running Neutron bootstrap container -------------------------- 43.34s\nneutron : Restart neutron-server container ----------------------------- 30.10s\nservice-ks-register : neutron | Granting user roles --------------------- 8.31s\nneutron : Copying over neutron.conf ------------------------------------- 6.44s\nservice-ks-register : neutron | Creating endpoints ---------------------- 6.39s\nneutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 3.90s\nservice-ks-register : neutron | Creating users -------------------------- 3.84s\nservice-ks-register : neutron | Creating services ----------------------- 3.76s\nneutron : Copying over ml2_conf.ini ------------------------------------- 3.58s\nservice-cert-copy : neutron | Copying over extra CA certificates -------- 3.51s\nneutron : Copying over config.json files for services ------------------- 3.39s\nservice-ks-register : neutron | Creating projects ----------------------- 3.13s\nservice-ks-register : neutron | Creating roles -------------------------- 3.09s\nservice-cert-copy : neutron | Copying over backend internal TLS key ----- 3.05s\nneutron : Copying over dnsmasq.conf ------------------------------------- 2.96s\nneutron : Check neutron containers -------------------------------------- 2.73s\nneutron : Copying over ssh key ------------------------------------------ 2.73s\nneutron : Ensuring config directories exist ----------------------------- 2.64s\nneutron : Copying over mlnx_agent.ini ----------------------------------- 2.63s\n\u001b[32m2026-03-26 03:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bd11feb7-7d46-47e4-ae00-c4b2147a1148 (nova) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bd11feb7-7d46-47e4-ae00-c4b2147a1148 (nova) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on OpenStack release] **********************************\nThursday 26 March 2026  03:15:18 +0000 (0:00:00.306)       0:00:00.306 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:15:19 +0000 (0:00:00.906)       0:00:01.213 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:15:20 +0000 (0:00:00.958)       0:00:02.172 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Bootstrap nova API databases] ********************************************\n\nTASK [Bootstrap deploy] ********************************************************\nThursday 26 March 2026  03:15:20 +0000 (0:00:00.860)       0:00:03.033 ******** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Creating Nova databases] ******************************************\nThursday 26 March 2026  03:15:21 +0000 (0:00:00.851)       0:00:03.884 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m\n\nTASK [nova : Creating Nova databases user and setting permissions] *************\nThursday 26 March 2026  03:15:25 +0000 (0:00:04.046)       0:00:07.930 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nThursday 26 March 2026  03:15:29 +0000 (0:00:04.101)       0:00:12.032 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over config.json files for nova-api-bootstrap] ************\nThursday 26 March 2026  03:15:30 +0000 (0:00:00.681)       0:00:12.714 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************\nThursday 26 March 2026  03:15:31 +0000 (0:00:01.307)       0:00:14.021 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nThursday 26 March 2026  03:15:34 +0000 (0:00:02.696)       0:00:16.718 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nThursday 26 March 2026  03:15:34 +0000 (0:00:00.330)       0:00:17.048 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Create cell0 mappings] ********************************************\nThursday 26 March 2026  03:16:07 +0000 (0:00:32.739)       0:00:49.788 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nThursday 26 March 2026  03:16:22 +0000 (0:00:14.467)       0:01:04.255 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nThursday 26 March 2026  03:16:33 +0000 (0:00:11.499)       0:01:15.754 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Update cell0 mappings] ********************************************\nThursday 26 March 2026  03:16:34 +0000 (0:00:00.705)       0:01:16.459 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nThursday 26 March 2026  03:16:34 +0000 (0:00:00.511)       0:01:16.970 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nThursday 26 March 2026  03:16:35 +0000 (0:00:00.753)       0:01:17.724 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nThursday 26 March 2026  03:16:52 +0000 (0:00:17.279)       0:01:35.004 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Bootstrap nova cell databases] *******************************************\n\nTASK [Bootstrap deploy] ********************************************************\nThursday 26 March 2026  03:16:53 +0000 (0:00:00.364)       0:01:35.368 ******** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database] *********************************\nThursday 26 March 2026  03:16:54 +0000 (0:00:00.879)       0:01:36.248 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database user and setting permissions] ****\nThursday 26 March 2026  03:16:56 +0000 (0:00:02.001)       0:01:38.250 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nThursday 26 March 2026  03:16:58 +0000 (0:00:02.058)       0:01:40.309 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nThursday 26 March 2026  03:16:58 +0000 (0:00:00.579)       0:01:40.888 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nThursday 26 March 2026  03:17:06 +0000 (0:00:07.383)       0:01:48.271 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nThursday 26 March 2026  03:17:06 +0000 (0:00:00.415)       0:01:48.686 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nThursday 26 March 2026  03:17:08 +0000 (0:00:01.576)       0:01:50.262 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******\nThursday 26 March 2026  03:17:08 +0000 (0:00:00.527)       0:01:50.790 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************\nThursday 26 March 2026  03:17:09 +0000 (0:00:01.006)       0:01:51.796 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Running Nova cell bootstrap container] ***********************\nThursday 26 March 2026  03:17:12 +0000 (0:00:02.636)       0:01:54.433 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nThursday 26 March 2026  03:17:34 +0000 (0:00:21.755)       0:02:16.188 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nThursday 26 March 2026  03:17:46 +0000 (0:00:12.347)       0:02:28.535 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Create cell] *************************************************\nThursday 26 March 2026  03:17:47 +0000 (0:00:01.198)       0:02:29.734 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Update cell] *************************************************\nThursday 26 March 2026  03:18:00 +0000 (0:00:12.613)       0:02:42.347 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nThursday 26 March 2026  03:18:01 +0000 (0:00:01.116)       0:02:43.463 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova] *********************************************************\n\nTASK [nova : include_tasks] ****************************************************\nThursday 26 March 2026  03:18:01 +0000 (0:00:00.339)       0:02:43.803 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : nova | Creating services] **************************\nThursday 26 March 2026  03:18:02 +0000 (0:00:00.809)       0:02:44.612 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m\n\nTASK [service-ks-register : nova | Creating endpoints] *************************\nThursday 26 March 2026  03:18:05 +0000 (0:00:03.184)       0:02:47.797 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m\n\nTASK [service-ks-register : nova | Creating projects] **************************\nThursday 26 March 2026  03:18:11 +0000 (0:00:06.199)       0:02:53.997 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating users] *****************************\nThursday 26 March 2026  03:18:15 +0000 (0:00:03.242)       0:02:57.240 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating roles] *****************************\nThursday 26 March 2026  03:18:18 +0000 (0:00:03.849)       0:03:01.089 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : nova | Granting user roles] ************************\nThursday 26 March 2026  03:18:22 +0000 (0:00:03.153)       0:03:04.243 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nThursday 26 March 2026  03:18:29 +0000 (0:00:07.345)       0:03:11.588 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Check if policies shall be overwritten] ***************************\nThursday 26 March 2026  03:18:30 +0000 (0:00:01.320)       0:03:12.909 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : Set nova policy file] *********************************************\nThursday 26 March 2026  03:18:30 +0000 (0:00:00.157)       0:03:13.067 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check for vendordata file] ****************************************\nThursday 26 March 2026  03:18:31 +0000 (0:00:00.341)       0:03:13.408 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [nova : Set vendordata file path] *****************************************\nThursday 26 March 2026  03:18:31 +0000 (0:00:00.688)       0:03:14.097 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nThursday 26 March 2026  03:18:32 +0000 (0:00:00.570)       0:03:14.668 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nThursday 26 March 2026  03:18:33 +0000 (0:00:00.595)       0:03:15.263 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:18:35 +0000 (0:00:02.389)       0:03:17.653 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nThursday 26 March 2026  03:18:36 +0000 (0:00:00.900)       0:03:18.554 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over config.json files for services] **********************\nThursday 26 March 2026  03:18:37 +0000 (0:00:00.859)       0:03:19.413 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over nova.conf] *******************************************\nThursday 26 March 2026  03:18:39 +0000 (0:00:02.478)       0:03:21.892 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over existing policy file] ********************************\nThursday 26 March 2026  03:18:45 +0000 (0:00:06.206)       0:03:28.098 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-api-wsgi.conf] **********************************\nThursday 26 March 2026  03:18:46 +0000 (0:00:00.622)       0:03:28.721 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over vendordata file] *************************************\nThursday 26 March 2026  03:18:48 +0000 (0:00:01.647)       0:03:30.368 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check nova containers] ********************************************\nThursday 26 March 2026  03:18:48 +0000 (0:00:00.342)       0:03:30.710 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Flush handlers] ***************************************************\nThursday 26 March 2026  03:18:50 +0000 (0:00:02.115)       0:03:32.826 ******** \n\nTASK [nova : Flush handlers] ***************************************************\nThursday 26 March 2026  03:18:50 +0000 (0:00:00.150)       0:03:32.977 ******** \n\nTASK [nova : Flush handlers] ***************************************************\nThursday 26 March 2026  03:18:50 +0000 (0:00:00.139)       0:03:33.117 ******** \n\nRUNNING HANDLER [nova : Restart nova-scheduler container] **********************\nThursday 26 March 2026  03:18:51 +0000 (0:00:00.148)       0:03:33.265 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-api container] ****************************\nThursday 26 March 2026  03:19:15 +0000 (0:00:23.876)       0:03:57.142 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role nova-cell] ****************************************************\n\nTASK [nova-cell : include_tasks] ***********************************************\nThursday 26 March 2026  03:19:20 +0000 (0:00:05.424)       0:04:02.566 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nThursday 26 March 2026  03:19:21 +0000 (0:00:01.330)       0:04:03.896 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Load and persist br_netfilter module] ************************************\nThursday 26 March 2026  03:19:22 +0000 (0:00:00.950)       0:04:04.847 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nThursday 26 March 2026  03:19:23 +0000 (0:00:00.895)       0:04:05.742 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nThursday 26 March 2026  03:19:24 +0000 (0:00:00.866)       0:04:06.609 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nThursday 26 March 2026  03:19:25 +0000 (0:00:01.247)       0:04:07.856 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************\nThursday 26 March 2026  03:19:26 +0000 (0:00:00.604)       0:04:08.461 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\nTASK [nova-cell : Install udev kolla kvm rules] ********************************\nThursday 26 March 2026  03:19:28 +0000 (0:00:02.015)       0:04:10.477 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Mask qemu-kvm service] ***************************************\nThursday 26 March 2026  03:19:29 +0000 (0:00:01.311)       0:04:11.789 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nThursday 26 March 2026  03:19:31 +0000 (0:00:01.830)       0:04:13.619 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nThursday 26 March 2026  03:19:33 +0000 (0:00:02.391)       0:04:16.011 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nThursday 26 March 2026  03:19:35 +0000 (0:00:01.483)       0:04:17.495 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:19:39 +0000 (0:00:03.709)       0:04:21.204 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nThursday 26 March 2026  03:19:40 +0000 (0:00:01.779)       0:04:22.983 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nThursday 26 March 2026  03:19:43 +0000 (0:00:02.222)       0:04:25.205 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check nova keyring file] *************************************\nThursday 26 March 2026  03:19:44 +0000 (0:00:01.124)       0:04:26.330 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Check cinder keyring file] ***********************************\nThursday 26 March 2026  03:19:45 +0000 (0:00:01.156)       0:04:27.487 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Extract nova key from file] **********************************\nThursday 26 March 2026  03:19:46 +0000 (0:00:00.973)       0:04:28.461 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Extract cinder key from file] ********************************\nThursday 26 March 2026  03:19:46 +0000 (0:00:00.558)       0:04:29.020 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph nova keyring file] ****************************\nThursday 26 March 2026  03:19:47 +0000 (0:00:00.540)       0:04:29.560 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph cinder keyring file] **************************\nThursday 26 March 2026  03:19:48 +0000 (0:00:01.434)       0:04:30.995 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf] *****************************************\nThursday 26 March 2026  03:19:50 +0000 (0:00:01.238)       0:04:32.233 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m\n\nTASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************\nThursday 26 March 2026  03:19:54 +0000 (0:00:03.905)       0:04:36.138 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************\nThursday 26 March 2026  03:19:54 +0000 (0:00:00.346)       0:04:36.484 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring libvirt secrets directory exists] *******************\nThursday 26 March 2026  03:19:54 +0000 (0:00:00.555)       0:04:37.040 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Pushing nova secret xml for libvirt] *************************\nThursday 26 March 2026  03:19:56 +0000 (0:00:01.266)       0:04:38.306 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\nTASK [nova-cell : Pushing secrets key for libvirt] *****************************\nThursday 26 March 2026  03:19:59 +0000 (0:00:03.341)       0:04:41.648 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Check if policies shall be overwritten] **********************\nThursday 26 March 2026  03:20:02 +0000 (0:00:03.453)       0:04:45.101 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Set nova policy file] ****************************************\nThursday 26 March 2026  03:20:03 +0000 (0:00:00.149)       0:04:45.250 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check for vendordata file] ***********************************\nThursday 26 March 2026  03:20:04 +0000 (0:00:00.932)       0:04:46.183 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\nTASK [nova-cell : Set vendordata file path] ************************************\nThursday 26 March 2026  03:20:04 +0000 (0:00:00.737)       0:04:46.921 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for services] *****************\nThursday 26 March 2026  03:20:05 +0000 (0:00:00.852)       0:04:47.773 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf] **************************************\nThursday 26 March 2026  03:20:09 +0000 (0:00:03.846)       0:04:51.619 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over Nova compute provider config] *******************\nThursday 26 March 2026  03:20:16 +0000 (0:00:06.982)       0:04:58.601 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt configuration] **************************\nThursday 26 March 2026  03:20:17 +0000 (0:00:01.365)       0:04:59.966 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\nTASK [nova-cell : Copying over libvirt TLS keys] *******************************\nThursday 26 March 2026  03:20:21 +0000 (0:00:03.720)       0:05:03.687 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt SASL configuration] *********************\nThursday 26 March 2026  03:20:22 +0000 (0:00:00.637)       0:05:04.324 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\nTASK [nova-cell : Copying files for nova-ssh] **********************************\nThursday 26 March 2026  03:20:27 +0000 (0:00:05.383)       0:05:09.708 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [nova-cell : Copying VMware vCenter CA file] ******************************\nThursday 26 March 2026  03:20:35 +0000 (0:00:07.669)       0:05:17.377 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying 'release' file for nova_compute] *********************\nThursday 26 March 2026  03:20:36 +0000 (0:00:00.839)       0:05:18.217 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************\nThursday 26 March 2026  03:20:36 +0000 (0:00:00.654)       0:05:18.871 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copying over existing policy file] ***************************\nThursday 26 March 2026  03:20:38 +0000 (0:00:02.172)       0:05:21.043 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Copying over vendordata file to containers] ******************\nThursday 26 March 2026  03:20:40 +0000 (0:00:01.563)       0:05:22.607 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check nova-cell containers] **********************************\nThursday 26 March 2026  03:20:41 +0000 (0:00:00.968)       0:05:23.576 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nThursday 26 March 2026  03:20:44 +0000 (0:00:02.706)       0:05:26.282 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:44 +0000 (0:00:00.830)       0:05:27.113 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:45 +0000 (0:00:00.169)       0:05:27.282 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:45 +0000 (0:00:00.149)       0:05:27.432 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:45 +0000 (0:00:00.154)       0:05:27.587 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:45 +0000 (0:00:00.148)       0:05:27.735 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nThursday 26 March 2026  03:20:45 +0000 (0:00:00.324)       0:05:28.060 ******** \n\nRUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************\nThursday 26 March 2026  03:20:46 +0000 (0:00:00.157)       0:05:28.218 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************\nThursday 26 March 2026  03:20:53 +0000 (0:00:07.475)       0:05:35.693 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************\nThursday 26 March 2026  03:21:09 +0000 (0:00:15.667)       0:05:51.361 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************\nThursday 26 March 2026  03:21:36 +0000 (0:00:27.431)       0:06:18.792 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************\nThursday 26 March 2026  03:22:19 +0000 (0:00:43.096)       0:07:01.889 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************\nThursday 26 March 2026  03:22:20 +0000 (0:00:00.751)       0:07:02.640 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-compute container] *******************\nThursday 26 March 2026  03:22:21 +0000 (0:00:00.848)       0:07:03.489 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***\nThursday 26 March 2026  03:22:47 +0000 (0:00:25.699)       0:07:29.189 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Waiting for nova-compute services to register themselves] ****\nThursday 26 March 2026  03:22:47 +0000 (0:00:00.148)       0:07:29.337 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if nova-compute service failed to register] *************\nThursday 26 March 2026  03:23:10 +0000 (0:00:22.999)       0:07:52.337 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Include discover_computes.yml] *******************************\nThursday 26 March 2026  03:23:20 +0000 (0:00:10.321)       0:08:02.658 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-4\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nThursday 26 March 2026  03:23:25 +0000 (0:00:04.471)       0:08:07.130 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nThursday 26 March 2026  03:23:37 +0000 (0:00:12.681)       0:08:19.811 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if cell settings not found] *****************************\nThursday 26 March 2026  03:23:39 +0000 (0:00:01.569)       0:08:21.381 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Discover nova hosts] *****************************************\nThursday 26 March 2026  03:23:41 +0000 (0:00:01.801)       0:08:23.182 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************\nThursday 26 March 2026  03:23:52 +0000 (0:00:11.180)       0:08:34.362 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Refresh nova scheduler cell cache] ***************************************\n\nTASK [nova : Refresh cell cache in nova scheduler] *****************************\nThursday 26 March 2026  03:23:54 +0000 (0:00:01.936)       0:08:36.298 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova super conductor services] *****************************\n\nTASK [nova : Reload nova super conductor services to remove RPC version pin] ***\nThursday 26 March 2026  03:23:55 +0000 (0:00:00.954)       0:08:37.253 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload Nova cell services] ***********************************************\n\nTASK [nova-cell : Reload nova cell services to remove RPC version cap] *********\nThursday 26 March 2026  03:23:55 +0000 (0:00:00.770)       0:08:38.023 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova API services] *****************************************\n\nTASK [nova : Reload nova API services to remove RPC version pin] ***************\nThursday 26 March 2026  03:23:57 +0000 (0:00:01.448)       0:08:39.472 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Run Nova API online data migrations] *************************************\n\nTASK [nova : Run Nova API online database migrations] **************************\nThursday 26 March 2026  03:23:58 +0000 (0:00:00.647)       0:08:40.119 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Run Nova cell online data migrations] ************************************\n\nTASK [nova-cell : Run Nova cell online database migrations] ********************\nThursday 26 March 2026  03:23:58 +0000 (0:00:00.986)       0:08:41.106 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=44  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:23:59 +0000 (0:00:00.495)       0:08:41.601 ******** \n=============================================================================== \nnova-cell : Restart nova-libvirt container ----------------------------- 43.10s\nnova : Running Nova API bootstrap container ---------------------------- 32.74s\nnova-cell : Restart nova-ssh container --------------------------------- 27.43s\nnova-cell : Restart nova-compute container ----------------------------- 25.70s\nnova : Restart nova-scheduler container -------------------------------- 23.88s\nnova-cell : Waiting for nova-compute services to register themselves --- 23.00s\nnova-cell : Running Nova cell bootstrap container ---------------------- 21.76s\nnova : Running Nova API bootstrap container ---------------------------- 17.28s\nnova-cell : Restart nova-novncproxy container -------------------------- 15.67s\nnova : Create cell0 mappings ------------------------------------------- 14.47s\nnova-cell : Get a list of existing cells ------------------------------- 12.68s\nnova-cell : Create cell ------------------------------------------------ 12.61s\nnova-cell : Get a list of existing cells ------------------------------- 12.35s\nnova-cell : Get a list of existing cells ------------------------------- 11.50s\nnova-cell : Discover nova hosts ---------------------------------------- 11.18s\nnova-cell : Fail if nova-compute service failed to register ------------ 10.32s\nnova-cell : Copying files for nova-ssh ---------------------------------- 7.67s\nnova-cell : Restart nova-conductor container ---------------------------- 7.48s\nservice-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 7.38s\nservice-ks-register : nova | Granting user roles ------------------------ 7.35s\n\u001b[32m2026-03-26 03:24:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f721f98e-b0b6-4aeb-a88b-3394e455017a (horizon) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:24:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f721f98e-b0b6-4aeb-a88b-3394e455017a (horizon) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:24:06 +0000 (0:00:00.301)       0:00:00.301 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:24:07 +0000 (0:00:00.356)       0:00:00.657 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m\n\nPLAY [Apply role horizon] ******************************************************\n\nTASK [horizon : include_tasks] *************************************************\nThursday 26 March 2026  03:24:07 +0000 (0:00:00.454)       0:00:01.112 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Ensuring config directories exist] *****************************\nThursday 26 March 2026  03:24:08 +0000 (0:00:00.535)       0:00:01.647 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : Set empty custom policy] ***************************************\nThursday 26 March 2026  03:24:09 +0000 (0:00:01.204)       0:00:02.852 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nThursday 26 March 2026  03:24:09 +0000 (0:00:00.538)       0:00:03.390 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:10 +0000 (0:00:00.825)       0:00:04.216 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:11 +0000 (0:00:00.371)       0:00:04.588 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:11 +0000 (0:00:00.330)       0:00:04.919 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:11 +0000 (0:00:00.325)       0:00:05.244 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:12 +0000 (0:00:00.344)       0:00:05.589 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:12 +0000 (0:00:00.149)       0:00:05.738 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:12 +0000 (0:00:00.341)       0:00:06.080 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:13 +0000 (0:00:00.540)       0:00:06.620 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:13 +0000 (0:00:00.134)       0:00:06.755 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:13 +0000 (0:00:00.332)       0:00:07.087 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:13 +0000 (0:00:00.325)       0:00:07.413 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:14 +0000 (0:00:00.148)       0:00:07.562 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:14 +0000 (0:00:00.568)       0:00:08.131 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:15 +0000 (0:00:00.343)       0:00:08.475 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:15 +0000 (0:00:00.143)       0:00:08.618 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:15 +0000 (0:00:00.335)       0:00:08.953 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:15 +0000 (0:00:00.350)       0:00:09.303 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:16 +0000 (0:00:00.334)       0:00:09.638 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:16 +0000 (0:00:00.358)       0:00:09.997 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:16 +0000 (0:00:00.353)       0:00:10.350 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:17 +0000 (0:00:00.165)       0:00:10.516 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:17 +0000 (0:00:00.323)       0:00:10.840 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:17 +0000 (0:00:00.587)       0:00:11.427 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:18 +0000 (0:00:00.153)       0:00:11.581 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:18 +0000 (0:00:00.374)       0:00:11.955 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:18 +0000 (0:00:00.395)       0:00:12.351 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:19 +0000 (0:00:00.129)       0:00:12.480 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nThursday 26 March 2026  03:24:19 +0000 (0:00:00.620)       0:00:13.100 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:24:20 +0000 (0:00:00.389)       0:00:13.490 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nThursday 26 March 2026  03:24:20 +0000 (0:00:00.156)       0:00:13.647 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over config.json files for services] *******************\nThursday 26 March 2026  03:24:20 +0000 (0:00:00.340)       0:00:13.987 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [horizon : Copying over horizon.conf] *************************************\nThursday 26 March 2026  03:24:22 +0000 (0:00:02.095)       0:00:16.082 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\nTASK [horizon : Copying over kolla-settings.py] ********************************\nThursday 26 March 2026  03:24:24 +0000 (0:00:02.088)       0:00:18.171 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over custom-settings.py] *******************************\nThursday 26 March 2026  03:24:26 +0000 (0:00:01.882)       0:00:20.053 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over existing policy file] *****************************\nThursday 26 March 2026  03:24:28 +0000 (0:00:01.562)       0:00:21.616 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over custom themes] ************************************\nThursday 26 March 2026  03:24:28 +0000 (0:00:00.566)       0:00:22.182 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nThursday 26 March 2026  03:24:29 +0000 (0:00:00.294)       0:00:22.477 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over extra CA certificates] ********\nThursday 26 March 2026  03:24:29 +0000 (0:00:00.652)       0:00:23.129 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:24:31 +0000 (0:00:02.119)       0:00:25.249 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****\nThursday 26 March 2026  03:24:32 +0000 (0:00:00.667)       0:00:25.917 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Deploy horizon container] **************************************\nThursday 26 March 2026  03:24:33 +0000 (0:00:00.894)       0:00:26.811 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nThursday 26 March 2026  03:24:35 +0000 (0:00:01.794)       0:00:28.606 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nThursday 26 March 2026  03:24:35 +0000 (0:00:00.325)       0:00:28.931 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Creating Horizon database] *************************************\nThursday 26 March 2026  03:24:36 +0000 (0:00:00.600)       0:00:29.532 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Creating Horizon database user and setting permissions] ********\nThursday 26 March 2026  03:24:38 +0000 (0:00:02.347)       0:00:31.879 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Running Horizon bootstrap container] ***************************\nThursday 26 March 2026  03:24:41 +0000 (0:00:02.704)       0:00:34.583 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Flush handlers] ************************************************\nThursday 26 March 2026  03:24:58 +0000 (0:00:17.350)       0:00:51.934 ******** \n\nTASK [horizon : Flush handlers] ************************************************\nThursday 26 March 2026  03:24:58 +0000 (0:00:00.087)       0:00:52.022 ******** \n\nTASK [horizon : Flush handlers] ************************************************\nThursday 26 March 2026  03:24:58 +0000 (0:00:00.066)       0:00:52.089 ******** \n\nRUNNING HANDLER [horizon : Restart horizon container] **************************\nThursday 26 March 2026  03:24:58 +0000 (0:00:00.082)       0:00:52.171 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:25:21 +0000 (0:00:22.557)       0:01:14.729 ******** \n=============================================================================== \nhorizon : Restart horizon container ------------------------------------ 22.56s\nhorizon : Running Horizon bootstrap container -------------------------- 17.35s\nhorizon : Creating Horizon database user and setting permissions -------- 2.70s\nhorizon : Creating Horizon database ------------------------------------- 2.35s\nservice-cert-copy : horizon | Copying over extra CA certificates -------- 2.12s\nhorizon : Copying over config.json files for services ------------------- 2.10s\nhorizon : Copying over horizon.conf ------------------------------------- 2.09s\nhorizon : Copying over kolla-settings.py -------------------------------- 1.88s\nhorizon : Deploy horizon container -------------------------------------- 1.79s\nhorizon : Copying over custom-settings.py ------------------------------- 1.56s\nhorizon : Ensuring config directories exist ----------------------------- 1.20s\nservice-cert-copy : horizon | Copying over backend internal TLS key ----- 0.89s\nhorizon : include_tasks ------------------------------------------------- 0.83s\nservice-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.67s\nhorizon : include_tasks ------------------------------------------------- 0.65s\nhorizon : Update custom policy file name -------------------------------- 0.62s\nhorizon : include_tasks ------------------------------------------------- 0.60s\nhorizon : Update policy file name --------------------------------------- 0.59s\nhorizon : Update custom policy file name -------------------------------- 0.57s\nhorizon : Copying over existing policy file ----------------------------- 0.57s\n\u001b[32m2026-03-26 03:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 845c9a2d-6656-4f8a-8038-0ee897175f20 (skyline) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 845c9a2d-6656-4f8a-8038-0ee897175f20 (skyline) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:25:28 +0000 (0:00:00.270)       0:00:00.270 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:25:28 +0000 (0:00:00.313)       0:00:00.584 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_skyline_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_skyline_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_skyline_True)\u001b[0m\n\nPLAY [Apply role skyline] ******************************************************\n\nTASK [skyline : include_tasks] *************************************************\nThursday 26 March 2026  03:25:29 +0000 (0:00:00.463)       0:00:01.048 ******** \n\u001b[0;36mincluded: /ansible/roles/skyline/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : skyline | Creating services] ***********************\nThursday 26 March 2026  03:25:29 +0000 (0:00:00.589)       0:00:01.637 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=skyline (panel))\u001b[0m\n\nTASK [service-ks-register : skyline | Creating endpoints] **********************\nThursday 26 March 2026  03:25:33 +0000 (0:00:03.454)       0:00:05.092 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> https://api-int.testbed.osism.xyz:9998 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> https://api.testbed.osism.xyz:9998 -> public)\u001b[0m\n\nTASK [service-ks-register : skyline | Creating projects] ***********************\nThursday 26 March 2026  03:25:40 +0000 (0:00:06.814)       0:00:11.906 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : skyline | Creating users] **************************\nThursday 26 March 2026  03:25:43 +0000 (0:00:02.984)       0:00:14.891 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> service)\u001b[0m\n\nTASK [service-ks-register : skyline | Creating roles] **************************\nThursday 26 March 2026  03:25:47 +0000 (0:00:04.064)       0:00:18.955 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : skyline | Granting user roles] *********************\nThursday 26 March 2026  03:25:50 +0000 (0:00:03.407)       0:00:22.363 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> service -> admin)\u001b[0m\n\nTASK [skyline : Ensuring config directories exist] *****************************\nThursday 26 March 2026  03:25:54 +0000 (0:00:03.844)       0:00:26.207 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [skyline : include_tasks] *************************************************\nThursday 26 March 2026  03:25:55 +0000 (0:00:01.339)       0:00:27.547 ******** \n\u001b[0;36mincluded: /ansible/roles/skyline/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : skyline | Copying over extra CA certificates] ********\nThursday 26 March 2026  03:25:56 +0000 (0:00:00.777)       0:00:28.324 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [service-cert-copy : skyline | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:25:59 +0000 (0:00:02.460)       0:00:30.785 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : skyline | Copying over backend internal TLS key] *****\nThursday 26 March 2026  03:25:59 +0000 (0:00:00.790)       0:00:31.576 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [skyline : Copying over skyline.yaml files for services] ******************\nThursday 26 March 2026  03:26:01 +0000 (0:00:01.340)       0:00:32.917 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [skyline : Copying over gunicorn.py files for services] *******************\nThursday 26 March 2026  03:26:03 +0000 (0:00:02.498)       0:00:35.415 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m\n\nTASK [skyline : Copying over nginx.conf files for services] ********************\nThursday 26 March 2026  03:26:05 +0000 (0:00:01.639)       0:00:37.055 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m\n\nTASK [skyline : Copying over config.json files for services] *******************\nThursday 26 March 2026  03:26:07 +0000 (0:00:02.137)       0:00:39.193 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [skyline : Copying over custom logos] *************************************\nThursday 26 March 2026  03:26:09 +0000 (0:00:02.307)       0:00:41.501 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [skyline : Check skyline container] ***************************************\nThursday 26 March 2026  03:26:10 +0000 (0:00:00.325)       0:00:41.826 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [skyline : Creating Skyline database] *************************************\nThursday 26 March 2026  03:26:12 +0000 (0:00:01.836)       0:00:43.663 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [skyline : Creating Skyline database user and setting permissions] ********\nThursday 26 March 2026  03:26:14 +0000 (0:00:02.288)       0:00:45.952 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [skyline : Running Skyline bootstrap container] ***************************\nThursday 26 March 2026  03:26:16 +0000 (0:00:02.406)       0:00:48.358 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [skyline : Flush handlers] ************************************************\nThursday 26 March 2026  03:26:24 +0000 (0:00:08.143)       0:00:56.502 ******** \n\nTASK [skyline : Flush handlers] ************************************************\nThursday 26 March 2026  03:26:24 +0000 (0:00:00.074)       0:00:56.576 ******** \n\nTASK [skyline : Flush handlers] ************************************************\nThursday 26 March 2026  03:26:25 +0000 (0:00:00.071)       0:00:56.648 ******** \n\nRUNNING HANDLER [skyline : Restart skyline-apiserver container] ****************\nThursday 26 March 2026  03:26:25 +0000 (0:00:00.083)       0:00:56.731 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [skyline : Restart skyline-console container] ******************\nThursday 26 March 2026  03:26:36 +0000 (0:00:11.142)       0:01:07.874 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:26:46 +0000 (0:00:09.775)       0:01:17.650 ******** \n=============================================================================== \nskyline : Restart skyline-apiserver container -------------------------- 11.14s\nskyline : Restart skyline-console container ----------------------------- 9.78s\nskyline : Running Skyline bootstrap container --------------------------- 8.14s\nservice-ks-register : skyline | Creating endpoints ---------------------- 6.81s\nservice-ks-register : skyline | Creating users -------------------------- 4.06s\nservice-ks-register : skyline | Granting user roles --------------------- 3.84s\nservice-ks-register : skyline | Creating services ----------------------- 3.45s\nservice-ks-register : skyline | Creating roles -------------------------- 3.41s\nservice-ks-register : skyline | Creating projects ----------------------- 2.98s\nskyline : Copying over skyline.yaml files for services ------------------ 2.50s\nservice-cert-copy : skyline | Copying over extra CA certificates -------- 2.46s\nskyline : Creating Skyline database user and setting permissions -------- 2.41s\nskyline : Copying over config.json files for services ------------------- 2.31s\nskyline : Creating Skyline database ------------------------------------- 2.29s\nskyline : Copying over nginx.conf files for services -------------------- 2.14s\nskyline : Check skyline container --------------------------------------- 1.84s\nskyline : Copying over gunicorn.py files for services ------------------- 1.64s\nservice-cert-copy : skyline | Copying over backend internal TLS key ----- 1.34s\nskyline : Ensuring config directories exist ----------------------------- 1.34s\nservice-cert-copy : skyline | Copying over backend internal TLS certificate --- 0.79s\n\u001b[32m2026-03-26 03:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06ad8f27-7892-4279-9b53-bb6de8a558f5 (glance) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06ad8f27-7892-4279-9b53-bb6de8a558f5 (glance) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:26:53 +0000 (0:00:00.298)       0:00:00.298 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:26:53 +0000 (0:00:00.436)       0:00:00.735 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m\n\nPLAY [Apply role glance] *******************************************************\n\nTASK [glance : include_tasks] **************************************************\nThursday 26 March 2026  03:26:54 +0000 (0:00:00.489)       0:00:01.225 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : glance | Creating services] ************************\nThursday 26 March 2026  03:26:54 +0000 (0:00:00.676)       0:00:01.901 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m\n\nTASK [service-ks-register : glance | Creating endpoints] ***********************\nThursday 26 March 2026  03:26:58 +0000 (0:00:03.400)       0:00:05.301 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m\n\nTASK [service-ks-register : glance | Creating projects] ************************\nThursday 26 March 2026  03:27:04 +0000 (0:00:06.406)       0:00:11.708 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating users] ***************************\nThursday 26 March 2026  03:27:07 +0000 (0:00:03.290)       0:00:14.999 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating roles] ***************************\nThursday 26 March 2026  03:27:12 +0000 (0:00:04.347)       0:00:19.346 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : glance | Granting user roles] **********************\nThursday 26 March 2026  03:27:15 +0000 (0:00:03.609)       0:00:22.956 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m\n\nTASK [glance : Ensuring config directories exist] ******************************\nThursday 26 March 2026  03:27:20 +0000 (0:00:04.187)       0:00:27.144 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nThursday 26 March 2026  03:27:23 +0000 (0:00:03.578)       0:00:30.723 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [glance : Ensuring glance service ceph config subdir exists] **************\nThursday 26 March 2026  03:27:24 +0000 (0:00:00.796)       0:00:31.519 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copy over multiple ceph configs for Glance] *********************\nThursday 26 March 2026  03:27:28 +0000 (0:00:03.690)       0:00:35.210 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Copy over ceph Glance keyrings] *********************************\nThursday 26 March 2026  03:27:29 +0000 (0:00:01.659)       0:00:36.869 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Ensuring config directory has correct owner and permission] *****\nThursday 26 March 2026  03:27:31 +0000 (0:00:01.448)       0:00:38.318 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check if policies shall be overwritten] *************************\nThursday 26 March 2026  03:27:32 +0000 (0:00:00.736)       0:00:39.055 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Set glance policy file] *****************************************\nThursday 26 March 2026  03:27:32 +0000 (0:00:00.139)       0:00:39.195 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nThursday 26 March 2026  03:27:32 +0000 (0:00:00.336)       0:00:39.531 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over extra CA certificates] *********\nThursday 26 March 2026  03:27:33 +0000 (0:00:00.910)       0:00:40.441 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:27:37 +0000 (0:00:04.121)       0:00:44.563 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS key] ******\nThursday 26 March 2026  03:27:40 +0000 (0:00:03.221)       0:00:47.784 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating TLS backend PEM File] **********************************\nThursday 26 March 2026  03:27:44 +0000 (0:00:03.615)       0:00:51.400 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over config.json files for services] ********************\nThursday 26 March 2026  03:27:47 +0000 (0:00:03.243)       0:00:54.643 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : Copying over glance-api.conf] ***********************************\nThursday 26 March 2026  03:27:51 +0000 (0:00:04.066)       0:00:58.710 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-cache.conf for glance_api] ******************\nThursday 26 March 2026  03:27:57 +0000 (0:00:05.749)       0:01:04.460 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-swift.conf for glance_api] ******************\nThursday 26 March 2026  03:28:01 +0000 (0:00:03.616)       0:01:08.077 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over glance-image-import.conf] **************************\nThursday 26 March 2026  03:28:04 +0000 (0:00:03.784)       0:01:11.861 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over property-protections-rules.conf] *******************\nThursday 26 March 2026  03:28:08 +0000 (0:00:03.647)       0:01:15.509 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over existing policy file] ******************************\nThursday 26 March 2026  03:28:12 +0000 (0:00:03.667)       0:01:19.177 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-haproxy-tls.cfg] ****************************\nThursday 26 March 2026  03:28:12 +0000 (0:00:00.582)       0:01:19.759 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostnqn' file for glance_api] ***********************\nThursday 26 March 2026  03:28:16 +0000 (0:00:03.340)       0:01:23.100 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check glance containers] ****************************************\nThursday 26 March 2026  03:28:20 +0000 (0:00:04.405)       0:01:27.506 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nThursday 26 March 2026  03:28:24 +0000 (0:00:03.983)       0:01:31.490 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating Glance database] ***************************************\nThursday 26 March 2026  03:28:25 +0000 (0:00:00.565)       0:01:32.056 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating Glance database user and setting permissions] **********\nThursday 26 March 2026  03:28:27 +0000 (0:00:02.210)       0:01:34.267 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Enable log_bin_trust_function_creators function] ****************\nThursday 26 March 2026  03:28:29 +0000 (0:00:02.446)       0:01:36.714 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Running Glance bootstrap container] *****************************\nThursday 26 March 2026  03:28:32 +0000 (0:00:02.423)       0:01:39.137 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Disable log_bin_trust_function_creators function] ***************\nThursday 26 March 2026  03:29:00 +0000 (0:00:28.855)       0:02:07.993 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Flush handlers] *************************************************\nThursday 26 March 2026  03:29:03 +0000 (0:00:02.219)       0:02:10.213 ******** \n\nTASK [glance : Flush handlers] *************************************************\nThursday 26 March 2026  03:29:03 +0000 (0:00:00.070)       0:02:10.283 ******** \n\nTASK [glance : Flush handlers] *************************************************\nThursday 26 March 2026  03:29:03 +0000 (0:00:00.071)       0:02:10.355 ******** \n\nRUNNING HANDLER [glance : Restart glance-api container] ************************\nThursday 26 March 2026  03:29:03 +0000 (0:00:00.076)       0:02:10.432 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:29:37 +0000 (0:00:33.781)       0:02:44.214 ******** \n=============================================================================== \nglance : Restart glance-api container ---------------------------------- 33.78s\nglance : Running Glance bootstrap container ---------------------------- 28.86s\nservice-ks-register : glance | Creating endpoints ----------------------- 6.41s\nglance : Copying over glance-api.conf ----------------------------------- 5.75s\nglance : Generating 'hostnqn' file for glance_api ----------------------- 4.41s\nservice-ks-register : glance | Creating users --------------------------- 4.35s\nservice-ks-register : glance | Granting user roles ---------------------- 4.19s\nservice-cert-copy : glance | Copying over extra CA certificates --------- 4.12s\nglance : Copying over config.json files for services -------------------- 4.07s\nglance : Check glance containers ---------------------------------------- 3.98s\nglance : Copying over glance-swift.conf for glance_api ------------------ 3.78s\nglance : Ensuring glance service ceph config subdir exists -------------- 3.69s\nglance : Copying over property-protections-rules.conf ------------------- 3.67s\nglance : Copying over glance-image-import.conf -------------------------- 3.65s\nglance : Copying over glance-cache.conf for glance_api ------------------ 3.62s\nservice-cert-copy : glance | Copying over backend internal TLS key ------ 3.62s\nservice-ks-register : glance | Creating roles --------------------------- 3.61s\nglance : Ensuring config directories exist ------------------------------ 3.58s\nservice-ks-register : glance | Creating services ------------------------ 3.40s\nglance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.34s\n\u001b[32m2026-03-26 03:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08eb036f-e85f-4744-96a6-01108fe8d3ed (cinder) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08eb036f-e85f-4744-96a6-01108fe8d3ed (cinder) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:29:44 +0000 (0:00:00.269)       0:00:00.269 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:29:44 +0000 (0:00:00.342)       0:00:00.612 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m\n\nPLAY [Apply role cinder] *******************************************************\n\nTASK [cinder : include_tasks] **************************************************\nThursday 26 March 2026  03:29:45 +0000 (0:00:00.465)       0:00:01.077 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : cinder | Creating services] ************************\nThursday 26 March 2026  03:29:45 +0000 (0:00:00.589)       0:00:01.667 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m\n\nTASK [service-ks-register : cinder | Creating endpoints] ***********************\nThursday 26 March 2026  03:29:49 +0000 (0:00:03.433)       0:00:05.100 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating projects] ************************\nThursday 26 March 2026  03:29:55 +0000 (0:00:06.462)       0:00:11.563 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating users] ***************************\nThursday 26 March 2026  03:29:58 +0000 (0:00:03.230)       0:00:14.793 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating roles] ***************************\nThursday 26 March 2026  03:30:02 +0000 (0:00:04.060)       0:00:18.853 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : cinder | Granting user roles] **********************\nThursday 26 March 2026  03:30:06 +0000 (0:00:03.240)       0:00:22.094 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m\n\nTASK [cinder : Ensuring config directories exist] ******************************\nThursday 26 March 2026  03:30:13 +0000 (0:00:07.226)       0:00:29.321 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nThursday 26 March 2026  03:30:15 +0000 (0:00:02.050)       0:00:31.371 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nThursday 26 March 2026  03:30:15 +0000 (0:00:00.520)       0:00:31.892 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [cinder : Ensuring cinder service ceph config subdirs exists] *************\nThursday 26 March 2026  03:30:16 +0000 (0:00:00.544)       0:00:32.437 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Copying over multiple ceph.conf for cinder services] ************\nThursday 26 March 2026  03:30:18 +0000 (0:00:01.624)       0:00:34.061 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************\nThursday 26 March 2026  03:30:21 +0000 (0:00:03.500)       0:00:37.561 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************\nThursday 26 March 2026  03:30:23 +0000 (0:00:01.630)       0:00:39.192 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\nTASK [cinder : Ensuring config directory has correct owner and permission] *****\nThursday 26 March 2026  03:30:25 +0000 (0:00:02.668)       0:00:41.860 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Check if policies shall be overwritten] *************************\nThursday 26 March 2026  03:30:26 +0000 (0:00:01.061)       0:00:42.921 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Set cinder policy file] *****************************************\nThursday 26 March 2026  03:30:27 +0000 (0:00:00.167)       0:00:43.089 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nThursday 26 March 2026  03:30:27 +0000 (0:00:00.510)       0:00:43.599 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over extra CA certificates] *********\nThursday 26 March 2026  03:30:28 +0000 (0:00:00.604)       0:00:44.204 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:30:32 +0000 (0:00:04.200)       0:00:48.405 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******\nThursday 26 March 2026  03:30:33 +0000 (0:00:00.986)       0:00:49.391 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over config.json files for services] ********************\nThursday 26 March 2026  03:30:34 +0000 (0:00:00.966)       0:00:50.357 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Copying over cinder-wsgi.conf] **********************************\nThursday 26 March 2026  03:30:38 +0000 (0:00:04.517)       0:00:54.875 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\nTASK [cinder : Copying over cinder.conf] ***************************************\nThursday 26 March 2026  03:30:40 +0000 (0:00:01.957)       0:00:56.833 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************\nThursday 26 March 2026  03:30:52 +0000 (0:00:11.264)       0:01:08.098 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [cinder : Copying over existing policy file] ******************************\nThursday 26 March 2026  03:30:53 +0000 (0:00:01.566)       0:01:09.664 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over nfs_shares files for cinder_volume] ****************\nThursday 26 March 2026  03:30:54 +0000 (0:00:00.950)       0:01:10.615 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Check cinder containers] ****************************************\nThursday 26 March 2026  03:30:55 +0000 (0:00:00.618)       0:01:11.233 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nThursday 26 March 2026  03:30:58 +0000 (0:00:02.975)       0:01:14.208 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Creating Cinder database] ***************************************\nThursday 26 March 2026  03:30:58 +0000 (0:00:00.307)       0:01:14.516 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Creating Cinder database user and setting permissions] **********\nThursday 26 March 2026  03:31:00 +0000 (0:00:02.141)       0:01:16.657 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Running Cinder bootstrap container] *****************************\nThursday 26 March 2026  03:31:02 +0000 (0:00:02.182)       0:01:18.840 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Flush handlers] *************************************************\nThursday 26 March 2026  03:31:23 +0000 (0:00:20.199)       0:01:39.039 ******** \n\nTASK [cinder : Flush handlers] *************************************************\nThursday 26 March 2026  03:31:23 +0000 (0:00:00.070)       0:01:39.110 ******** \n\nTASK [cinder : Flush handlers] *************************************************\nThursday 26 March 2026  03:31:23 +0000 (0:00:00.076)       0:01:39.187 ******** \n\nRUNNING HANDLER [cinder : Restart cinder-api container] ************************\nThursday 26 March 2026  03:31:23 +0000 (0:00:00.074)       0:01:39.261 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************\nThursday 26 March 2026  03:31:49 +0000 (0:00:26.597)       0:02:05.859 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-volume container] *********************\nThursday 26 March 2026  03:31:58 +0000 (0:00:08.377)       0:02:14.237 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-backup container] *********************\nThursday 26 March 2026  03:32:25 +0000 (0:00:27.727)       0:02:41.964 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***\nThursday 26 March 2026  03:32:37 +0000 (0:00:11.022)       0:02:52.986 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:32:37 +0000 (0:00:00.284)       0:02:53.271 ******** \n=============================================================================== \ncinder : Restart cinder-volume container ------------------------------- 27.73s\ncinder : Restart cinder-api container ---------------------------------- 26.60s\ncinder : Running Cinder bootstrap container ---------------------------- 20.20s\ncinder : Copying over cinder.conf -------------------------------------- 11.26s\ncinder : Restart cinder-backup container ------------------------------- 11.02s\ncinder : Restart cinder-scheduler container ----------------------------- 8.38s\nservice-ks-register : cinder | Granting user roles ---------------------- 7.23s\nservice-ks-register : cinder | Creating endpoints ----------------------- 6.46s\ncinder : Copying over config.json files for services -------------------- 4.52s\nservice-cert-copy : cinder | Copying over extra CA certificates --------- 4.20s\nservice-ks-register : cinder | Creating users --------------------------- 4.06s\ncinder : Copying over multiple ceph.conf for cinder services ------------ 3.50s\nservice-ks-register : cinder | Creating services ------------------------ 3.43s\nservice-ks-register : cinder | Creating roles --------------------------- 3.24s\nservice-ks-register : cinder | Creating projects ------------------------ 3.23s\ncinder : Check cinder containers ---------------------------------------- 2.98s\ncinder : Copy over Ceph keyring files for cinder-backup ----------------- 2.67s\ncinder : Creating Cinder database user and setting permissions ---------- 2.18s\ncinder : Creating Cinder database --------------------------------------- 2.14s\ncinder : Ensuring config directories exist ------------------------------ 2.05s\n\u001b[32m2026-03-26 03:32:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 68ea24c0-1a9f-4835-83c3-be9f58b464fd (barbican) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:32:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 68ea24c0-1a9f-4835-83c3-be9f58b464fd (barbican) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:32:44 +0000 (0:00:00.286)       0:00:00.286 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:32:44 +0000 (0:00:00.332)       0:00:00.618 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m\n\nPLAY [Apply role barbican] *****************************************************\n\nTASK [barbican : include_tasks] ************************************************\nThursday 26 March 2026  03:32:45 +0000 (0:00:00.496)       0:00:01.114 ******** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : barbican | Creating services] **********************\nThursday 26 March 2026  03:32:45 +0000 (0:00:00.582)       0:00:01.697 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m\n\nTASK [service-ks-register : barbican | Creating endpoints] *********************\nThursday 26 March 2026  03:32:49 +0000 (0:00:03.493)       0:00:05.190 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating projects] **********************\nThursday 26 March 2026  03:32:55 +0000 (0:00:06.389)       0:00:11.580 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating users] *************************\nThursday 26 March 2026  03:32:59 +0000 (0:00:03.211)       0:00:14.791 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating roles] *************************\nThursday 26 March 2026  03:33:03 +0000 (0:00:04.371)       0:00:19.162 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m\n\nTASK [service-ks-register : barbican | Granting user roles] ********************\nThursday 26 March 2026  03:33:19 +0000 (0:00:16.079)       0:00:35.242 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m\n\nTASK [barbican : Ensuring config directories exist] ****************************\nThursday 26 March 2026  03:33:23 +0000 (0:00:03.806)       0:00:39.049 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Ensuring vassals config directories exist] ********************\nThursday 26 March 2026  03:33:24 +0000 (0:00:01.624)       0:00:40.673 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m\n\nTASK [barbican : Check if policies shall be overwritten] ***********************\nThursday 26 March 2026  03:33:26 +0000 (0:00:01.236)       0:00:41.910 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Set barbican policy file] *************************************\nThursday 26 March 2026  03:33:26 +0000 (0:00:00.368)       0:00:42.278 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nThursday 26 March 2026  03:33:26 +0000 (0:00:00.326)       0:00:42.604 ******** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over extra CA certificates] *******\nThursday 26 March 2026  03:33:27 +0000 (0:00:00.581)       0:00:43.185 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:33:30 +0000 (0:00:03.472)       0:00:46.658 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****\nThursday 26 March 2026  03:33:31 +0000 (0:00:00.584)       0:00:47.242 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over config.json files for services] ******************\nThursday 26 March 2026  03:33:32 +0000 (0:00:00.828)       0:00:48.070 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over barbican-api.ini] ********************************\nThursday 26 March 2026  03:33:35 +0000 (0:00:03.616)       0:00:51.687 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Checking whether barbican-api-paste.ini file exists] **********\nThursday 26 March 2026  03:33:37 +0000 (0:00:01.625)       0:00:53.313 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [barbican : Copying over barbican-api-paste.ini] **************************\nThursday 26 March 2026  03:33:38 +0000 (0:00:01.031)       0:00:54.345 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over barbican.conf] ***********************************\nThursday 26 March 2026  03:33:39 +0000 (0:00:00.592)       0:00:54.938 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over existing policy file] ****************************\nThursday 26 March 2026  03:33:45 +0000 (0:00:06.794)       0:01:01.732 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Check barbican containers] ************************************\nThursday 26 March 2026  03:33:46 +0000 (0:00:00.856)       0:01:02.588 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nThursday 26 March 2026  03:33:49 +0000 (0:00:02.397)       0:01:04.986 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Creating barbican database] ***********************************\nThursday 26 March 2026  03:33:49 +0000 (0:00:00.328)       0:01:05.314 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Creating barbican database user and setting permissions] ******\nThursday 26 March 2026  03:33:51 +0000 (0:00:02.031)       0:01:07.345 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Running barbican bootstrap container] *************************\nThursday 26 March 2026  03:33:53 +0000 (0:00:02.220)       0:01:09.566 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Flush handlers] ***********************************************\nThursday 26 March 2026  03:34:06 +0000 (0:00:12.503)       0:01:22.069 ******** \n\nTASK [barbican : Flush handlers] ***********************************************\nThursday 26 March 2026  03:34:06 +0000 (0:00:00.073)       0:01:22.142 ******** \n\nTASK [barbican : Flush handlers] ***********************************************\nThursday 26 March 2026  03:34:06 +0000 (0:00:00.083)       0:01:22.226 ******** \n\nRUNNING HANDLER [barbican : Restart barbican-api container] ********************\nThursday 26 March 2026  03:34:06 +0000 (0:00:00.082)       0:01:22.309 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******\nThursday 26 March 2026  03:34:18 +0000 (0:00:11.526)       0:01:33.835 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-worker container] *****************\nThursday 26 March 2026  03:34:23 +0000 (0:00:05.071)       0:01:38.907 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:34:33 +0000 (0:00:10.439)       0:01:49.346 ******** \n=============================================================================== \nservice-ks-register : barbican | Creating roles ------------------------ 16.08s\nbarbican : Running barbican bootstrap container ------------------------ 12.50s\nbarbican : Restart barbican-api container ------------------------------ 11.53s\nbarbican : Restart barbican-worker container --------------------------- 10.44s\nbarbican : Copying over barbican.conf ----------------------------------- 6.79s\nservice-ks-register : barbican | Creating endpoints --------------------- 6.39s\nbarbican : Restart barbican-keystone-listener container ----------------- 5.07s\nservice-ks-register : barbican | Creating users ------------------------- 4.37s\nservice-ks-register : barbican | Granting user roles -------------------- 3.81s\nbarbican : Copying over config.json files for services ------------------ 3.62s\nservice-ks-register : barbican | Creating services ---------------------- 3.49s\nservice-cert-copy : barbican | Copying over extra CA certificates ------- 3.47s\nservice-ks-register : barbican | Creating projects ---------------------- 3.21s\nbarbican : Check barbican containers ------------------------------------ 2.40s\nbarbican : Creating barbican database user and setting permissions ------ 2.22s\nbarbican : Creating barbican database ----------------------------------- 2.03s\nbarbican : Copying over barbican-api.ini -------------------------------- 1.63s\nbarbican : Ensuring config directories exist ---------------------------- 1.62s\nbarbican : Ensuring vassals config directories exist -------------------- 1.24s\nbarbican : Checking whether barbican-api-paste.ini file exists ---------- 1.03s\n\u001b[32m2026-03-26 03:34:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9fd4a647-2541-4c2c-a7fe-e47d95f2c814 (designate) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:34:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9fd4a647-2541-4c2c-a7fe-e47d95f2c814 (designate) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:34:40 +0000 (0:00:00.295)       0:00:00.295 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:34:41 +0000 (0:00:00.332)       0:00:00.628 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m\n\nPLAY [Apply role designate] ****************************************************\n\nTASK [designate : include_tasks] ***********************************************\nThursday 26 March 2026  03:34:41 +0000 (0:00:00.485)       0:00:01.113 ******** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : designate | Creating services] *********************\nThursday 26 March 2026  03:34:42 +0000 (0:00:00.570)       0:00:01.684 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m\n\nTASK [service-ks-register : designate | Creating endpoints] ********************\nThursday 26 March 2026  03:34:45 +0000 (0:00:03.465)       0:00:05.149 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m\n\nTASK [service-ks-register : designate | Creating projects] *********************\nThursday 26 March 2026  03:34:52 +0000 (0:00:06.464)       0:00:11.614 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating users] ************************\nThursday 26 March 2026  03:34:55 +0000 (0:00:03.235)       0:00:14.850 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating roles] ************************\nThursday 26 March 2026  03:34:59 +0000 (0:00:04.052)       0:00:18.902 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : designate | Granting user roles] *******************\nThursday 26 March 2026  03:35:02 +0000 (0:00:03.286)       0:00:22.188 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m\n\nTASK [designate : Ensuring config directories exist] ***************************\nThursday 26 March 2026  03:35:06 +0000 (0:00:03.969)       0:00:26.158 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Check if policies shall be overwritten] **********************\nThursday 26 March 2026  03:35:09 +0000 (0:00:02.998)       0:00:29.157 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Set designate policy file] ***********************************\nThursday 26 March 2026  03:35:10 +0000 (0:00:00.146)       0:00:29.304 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nThursday 26 March 2026  03:35:10 +0000 (0:00:00.584)       0:00:29.888 ******** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over extra CA certificates] ******\nThursday 26 March 2026  03:35:11 +0000 (0:00:00.574)       0:00:30.462 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:35:17 +0000 (0:00:06.039)       0:00:36.501 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS key] ***\nThursday 26 March 2026  03:35:18 +0000 (0:00:01.058)       0:00:37.560 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [designate : Copying over config.json files for services] *****************\nThursday 26 March 2026  03:35:19 +0000 (0:00:01.214)       0:00:38.774 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over designate.conf] *********************************\nThursday 26 March 2026  03:35:25 +0000 (0:00:06.202)       0:00:44.976 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over pools.yaml] *************************************\nThursday 26 March 2026  03:35:40 +0000 (0:00:15.070)       0:01:00.046 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\nTASK [designate : Copying over named.conf] *************************************\nThursday 26 March 2026  03:35:44 +0000 (0:00:03.756)       0:01:03.802 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\nTASK [designate : Copying over rndc.conf] **************************************\nThursday 26 March 2026  03:35:47 +0000 (0:00:02.632)       0:01:06.435 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over rndc.key] ***************************************\nThursday 26 March 2026  03:35:50 +0000 (0:00:03.033)       0:01:09.469 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nThursday 26 March 2026  03:35:53 +0000 (0:00:02.845)       0:01:12.314 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over existing policy file] ***************************\nThursday 26 March 2026  03:35:53 +0000 (0:00:00.331)       0:01:12.646 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Check designate containers] **********************************\nThursday 26 March 2026  03:35:54 +0000 (0:00:00.800)       0:01:13.447 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nThursday 26 March 2026  03:35:58 +0000 (0:00:04.719)       0:01:18.166 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Creating Designate databases] ********************************\nThursday 26 March 2026  03:35:59 +0000 (0:00:00.339)       0:01:18.506 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m\n\nTASK [designate : Creating Designate databases user and setting permissions] ***\nThursday 26 March 2026  03:36:01 +0000 (0:00:02.148)       0:01:20.655 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m\n\nTASK [designate : Running Designate bootstrap container] ***********************\nThursday 26 March 2026  03:36:03 +0000 (0:00:02.342)       0:01:22.997 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [designate : Flush handlers] **********************************************\nThursday 26 March 2026  03:36:19 +0000 (0:00:16.271)       0:01:39.269 ******** \n\nTASK [designate : Flush handlers] **********************************************\nThursday 26 March 2026  03:36:20 +0000 (0:00:00.073)       0:01:39.342 ******** \n\nTASK [designate : Flush handlers] **********************************************\nThursday 26 March 2026  03:36:20 +0000 (0:00:00.074)       0:01:39.417 ******** \n\nRUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********\nThursday 26 March 2026  03:36:20 +0000 (0:00:00.072)       0:01:39.489 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-api container] ******************\nThursday 26 March 2026  03:36:28 +0000 (0:00:07.943)       0:01:47.432 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-central container] **************\nThursday 26 March 2026  03:36:38 +0000 (0:00:10.718)       0:01:58.151 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-producer container] *************\nThursday 26 March 2026  03:36:44 +0000 (0:00:05.939)       0:02:04.091 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-mdns container] *****************\nThursday 26 March 2026  03:36:50 +0000 (0:00:05.941)       0:02:10.032 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-worker container] ***************\nThursday 26 March 2026  03:36:56 +0000 (0:00:06.071)       0:02:16.104 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [designate : Non-destructive DNS pools update] ****************************\nThursday 26 March 2026  03:37:07 +0000 (0:00:11.022)       0:02:27.127 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:37:15 +0000 (0:00:07.390)       0:02:34.517 ******** \n=============================================================================== \ndesignate : Running Designate bootstrap container ---------------------- 16.27s\ndesignate : Copying over designate.conf -------------------------------- 15.07s\ndesignate : Restart designate-worker container ------------------------- 11.02s\ndesignate : Restart designate-api container ---------------------------- 10.72s\ndesignate : Restart designate-backend-bind9 container ------------------- 7.94s\ndesignate : Non-destructive DNS pools update ---------------------------- 7.39s\nservice-ks-register : designate | Creating endpoints -------------------- 6.46s\ndesignate : Copying over config.json files for services ----------------- 6.20s\ndesignate : Restart designate-mdns container ---------------------------- 6.07s\nservice-cert-copy : designate | Copying over extra CA certificates ------ 6.04s\ndesignate : Restart designate-producer container ------------------------ 5.94s\ndesignate : Restart designate-central container ------------------------- 5.94s\ndesignate : Check designate containers ---------------------------------- 4.72s\nservice-ks-register : designate | Creating users ------------------------ 4.05s\nservice-ks-register : designate | Granting user roles ------------------- 3.97s\ndesignate : Copying over pools.yaml ------------------------------------- 3.76s\nservice-ks-register : designate | Creating services --------------------- 3.47s\nservice-ks-register : designate | Creating roles ------------------------ 3.29s\nservice-ks-register : designate | Creating projects --------------------- 3.24s\ndesignate : Copying over rndc.conf -------------------------------------- 3.03s\n\u001b[32m2026-03-26 03:37:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 435223f9-b996-4394-82a5-4bcb72d6599b (octavia) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:37:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 435223f9-b996-4394-82a5-4bcb72d6599b (octavia) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:37:22 +0000 (0:00:00.271)       0:00:00.271 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:37:23 +0000 (0:00:00.307)       0:00:00.579 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m\n\nPLAY [Apply role octavia] ******************************************************\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:37:23 +0000 (0:00:00.461)       0:00:01.041 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : octavia | Creating services] ***********************\nThursday 26 March 2026  03:37:24 +0000 (0:00:00.579)       0:00:01.621 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m\n\nTASK [service-ks-register : octavia | Creating endpoints] **********************\nThursday 26 March 2026  03:37:27 +0000 (0:00:03.307)       0:00:04.928 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating projects] ***********************\nThursday 26 March 2026  03:37:34 +0000 (0:00:06.535)       0:00:11.464 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating users] **************************\nThursday 26 March 2026  03:37:37 +0000 (0:00:03.320)       0:00:14.785 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating roles] **************************\nThursday 26 March 2026  03:37:45 +0000 (0:00:08.315)       0:00:23.100 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : octavia | Granting user roles] *********************\nThursday 26 March 2026  03:37:49 +0000 (0:00:03.289)       0:00:26.390 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\nTASK [octavia : Adding octavia related roles] **********************************\nThursday 26 March 2026  03:37:56 +0000 (0:00:07.620)       0:00:34.010 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:38:12 +0000 (0:00:16.050)       0:00:50.060 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create amphora flavor] *****************************************\nThursday 26 March 2026  03:38:13 +0000 (0:00:00.780)       0:00:50.840 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create nova keypair for amphora] *******************************\nThursday 26 March 2026  03:38:18 +0000 (0:00:05.038)       0:00:55.879 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nThursday 26 March 2026  03:38:23 +0000 (0:00:04.454)       0:01:00.333 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create security groups for octavia] ****************************\nThursday 26 March 2026  03:38:26 +0000 (0:00:03.281)       0:01:03.615 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Add rules for security groups] *********************************\nThursday 26 March 2026  03:38:36 +0000 (0:00:10.054)       0:01:13.669 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m\n\nTASK [octavia : Create loadbalancer management network] ************************\nThursday 26 March 2026  03:38:53 +0000 (0:00:16.922)       0:01:30.592 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management subnet] *************************\nThursday 26 March 2026  03:38:57 +0000 (0:00:04.607)       0:01:35.199 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management router for IPv6] ****************\nThursday 26 March 2026  03:39:03 +0000 (0:00:05.523)       0:01:40.723 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update loadbalancer management subnet] *************************\nThursday 26 March 2026  03:39:03 +0000 (0:00:00.221)       0:01:40.944 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:39:08 +0000 (0:00:04.459)       0:01:45.403 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create ports for Octavia health-manager nodes] *****************\nThursday 26 March 2026  03:39:09 +0000 (0:00:01.158)       0:01:46.561 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update Octavia health manager port host_id] ********************\nThursday 26 March 2026  03:39:14 +0000 (0:00:05.629)       0:01:52.191 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Add Octavia port to openvswitch br-int] ************************\nThursday 26 March 2026  03:39:19 +0000 (0:00:04.313)       0:01:56.505 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Install isc-dhcp-client package] *******************************\nThursday 26 March 2026  03:39:20 +0000 (0:00:01.047)       0:01:57.552 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia dhclient conf] **********************************\nThursday 26 March 2026  03:39:21 +0000 (0:00:01.716)       0:01:59.268 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia-interface service] ******************************\nThursday 26 March 2026  03:39:23 +0000 (0:00:01.276)       0:02:00.545 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Restart octavia-interface.service if required] *****************\nThursday 26 March 2026  03:39:24 +0000 (0:00:01.188)       0:02:01.734 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Enable and start octavia-interface.service] ********************\nThursday 26 March 2026  03:39:26 +0000 (0:00:02.205)       0:02:03.939 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Wait for interface ohm0 ip appear] *****************************\nThursday 26 March 2026  03:39:28 +0000 (0:00:01.515)       0:02:05.455 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Gather facts] **************************************************\nThursday 26 March 2026  03:39:28 +0000 (0:00:00.643)       0:02:06.099 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:39:31 +0000 (0:00:03.116)       0:02:09.216 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Get amphora flavor info] ***************************************\nThursday 26 March 2026  03:39:32 +0000 (0:00:00.575)       0:02:09.791 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nThursday 26 March 2026  03:39:36 +0000 (0:00:03.990)       0:02:13.781 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get security groups for octavia] *******************************\nThursday 26 March 2026  03:39:39 +0000 (0:00:03.294)       0:02:17.076 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Get loadbalancer management network] ***************************\nThursday 26 March 2026  03:39:47 +0000 (0:00:07.349)       0:02:24.425 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia resources facts] ***********************************\nThursday 26 March 2026  03:39:51 +0000 (0:00:04.216)       0:02:28.641 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Ensuring config directories exist] *****************************\nThursday 26 March 2026  03:39:51 +0000 (0:00:00.555)       0:02:29.197 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Check if policies shall be overwritten] ************************\nThursday 26 March 2026  03:39:54 +0000 (0:00:02.522)       0:02:31.719 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia policy file] ***************************************\nThursday 26 March 2026  03:39:54 +0000 (0:00:00.140)       0:02:31.860 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over existing policy file] *****************************\nThursday 26 March 2026  03:39:54 +0000 (0:00:00.311)       0:02:32.171 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:39:55 +0000 (0:00:00.704)       0:02:32.876 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over extra CA certificates] ********\nThursday 26 March 2026  03:39:56 +0000 (0:00:00.754)       0:02:33.631 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:40:01 +0000 (0:00:04.963)       0:02:38.594 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****\nThursday 26 March 2026  03:40:01 +0000 (0:00:00.668)       0:02:39.262 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over config.json files for services] *******************\nThursday 26 March 2026  03:40:03 +0000 (0:00:01.337)       0:02:40.600 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over octavia-wsgi.conf] ********************************\nThursday 26 March 2026  03:40:08 +0000 (0:00:05.318)       0:02:45.919 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\nTASK [octavia : Copying over octavia.conf] *************************************\nThursday 26 March 2026  03:40:10 +0000 (0:00:01.698)       0:02:47.617 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over Octavia SSH key] **********************************\nThursday 26 March 2026  03:40:27 +0000 (0:00:17.437)       0:03:05.055 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-worker] ******************\nThursday 26 March 2026  03:40:29 +0000 (0:00:01.902)       0:03:06.958 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-housekeeping] ************\nThursday 26 March 2026  03:40:34 +0000 (0:00:05.103)       0:03:12.061 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-health-manager] **********\nThursday 26 March 2026  03:40:39 +0000 (0:00:05.205)       0:03:17.266 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Check octavia containers] **************************************\nThursday 26 March 2026  03:40:45 +0000 (0:00:05.318)       0:03:22.585 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nThursday 26 March 2026  03:40:49 +0000 (0:00:04.089)       0:03:26.675 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Creating Octavia database] *************************************\nThursday 26 March 2026  03:40:49 +0000 (0:00:00.588)       0:03:27.263 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database] *************************\nThursday 26 March 2026  03:40:52 +0000 (0:00:02.237)       0:03:29.501 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia database user and setting permissions] ********\nThursday 26 March 2026  03:40:54 +0000 (0:00:02.128)       0:03:31.630 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database user and setting permissions] ***\nThursday 26 March 2026  03:40:56 +0000 (0:00:02.233)       0:03:33.863 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Running Octavia bootstrap container] ***************************\nThursday 26 March 2026  03:40:58 +0000 (0:00:02.189)       0:03:36.053 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Flush handlers] ************************************************\nThursday 26 March 2026  03:41:21 +0000 (0:00:23.142)       0:03:59.196 ******** \n\nTASK [octavia : Flush handlers] ************************************************\nThursday 26 March 2026  03:41:21 +0000 (0:00:00.073)       0:03:59.269 ******** \n\nTASK [octavia : Flush handlers] ************************************************\nThursday 26 March 2026  03:41:22 +0000 (0:00:00.070)       0:03:59.340 ******** \n\nRUNNING HANDLER [octavia : Restart octavia-api container] **********************\nThursday 26 March 2026  03:41:22 +0000 (0:00:00.071)       0:03:59.411 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************\nThursday 26 March 2026  03:41:39 +0000 (0:00:17.243)       0:04:16.654 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********\nThursday 26 March 2026  03:41:50 +0000 (0:00:11.439)       0:04:28.093 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************\nThursday 26 March 2026  03:42:01 +0000 (0:00:10.580)       0:04:38.674 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-worker container] *******************\nThursday 26 March 2026  03:42:11 +0000 (0:00:10.104)       0:04:48.778 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:42:17 +0000 (0:00:05.841)       0:04:54.619 ******** \n=============================================================================== \noctavia : Running Octavia bootstrap container -------------------------- 23.14s\noctavia : Copying over octavia.conf ------------------------------------ 17.44s\noctavia : Restart octavia-api container -------------------------------- 17.24s\noctavia : Add rules for security groups -------------------------------- 16.92s\noctavia : Adding octavia related roles --------------------------------- 16.05s\noctavia : Restart octavia-driver-agent container ----------------------- 11.44s\noctavia : Restart octavia-health-manager container --------------------- 10.58s\noctavia : Restart octavia-housekeeping container ----------------------- 10.10s\noctavia : Create security groups for octavia --------------------------- 10.05s\nservice-ks-register : octavia | Creating users -------------------------- 8.32s\nservice-ks-register : octavia | Granting user roles --------------------- 7.62s\noctavia : Get security groups for octavia ------------------------------- 7.35s\nservice-ks-register : octavia | Creating endpoints ---------------------- 6.54s\noctavia : Restart octavia-worker container ------------------------------ 5.84s\noctavia : Create ports for Octavia health-manager nodes ----------------- 5.63s\noctavia : Create loadbalancer management subnet ------------------------- 5.52s\noctavia : Copying certificate files for octavia-health-manager ---------- 5.32s\noctavia : Copying over config.json files for services ------------------- 5.32s\noctavia : Copying certificate files for octavia-housekeeping ------------ 5.21s\noctavia : Copying certificate files for octavia-worker ------------------ 5.10s\n\u001b[32m2026-03-26 03:42:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ced74339-9146-4768-949d-5ed80733d53c (ceilometer) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:42:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ced74339-9146-4768-949d-5ed80733d53c (ceilometer) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:42:24 +0000 (0:00:00.277)       0:00:00.277 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:42:25 +0000 (0:00:00.720)       0:00:00.998 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceilometer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceilometer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceilometer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceilometer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceilometer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceilometer_True)\u001b[0m\n\nPLAY [Apply role ceilometer] ***************************************************\n\nTASK [ceilometer : include_tasks] **********************************************\nThursday 26 March 2026  03:42:25 +0000 (0:00:00.636)       0:00:01.635 ******** \n\u001b[0;36mincluded: /ansible/roles/ceilometer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-ks-register : ceilometer | Creating services] ********************\nThursday 26 March 2026  03:42:27 +0000 (0:00:01.271)       0:00:02.906 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : ceilometer | Creating endpoints] *******************\nThursday 26 March 2026  03:42:27 +0000 (0:00:00.131)       0:00:03.038 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : ceilometer | Creating projects] ********************\nThursday 26 March 2026  03:42:27 +0000 (0:00:00.136)       0:00:03.174 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceilometer | Creating users] ***********************\nThursday 26 March 2026  03:42:31 +0000 (0:00:03.889)       0:00:07.064 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceilometer -> service)\u001b[0m\n\nTASK [service-ks-register : ceilometer | Creating roles] ***********************\nThursday 26 March 2026  03:42:35 +0000 (0:00:04.001)       0:00:11.065 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : ceilometer | Granting user roles] ******************\nThursday 26 March 2026  03:42:38 +0000 (0:00:03.231)       0:00:14.296 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceilometer -> service -> admin)\u001b[0m\n\nTASK [ceilometer : Associate the ResellerAdmin role and ceilometer user] *******\nThursday 26 March 2026  03:42:42 +0000 (0:00:04.305)       0:00:18.602 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceilometer : Ensuring config directories exist] **************************\nThursday 26 March 2026  03:42:42 +0000 (0:00:00.139)       0:00:18.742 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [ceilometer : Check if the folder for custom meter definitions exist] *****\nThursday 26 March 2026  03:42:44 +0000 (0:00:01.453)       0:00:20.196 ******** \n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [ceilometer : Set variable that indicates if we have a folder for custom meter YAML files] ***\nThursday 26 March 2026  03:42:46 +0000 (0:00:01.683)       0:00:21.880 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Find all *.yaml files in custom meter definitions folder (if the folder exist)] ***\nThursday 26 March 2026  03:42:46 +0000 (0:00:00.623)       0:00:22.503 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Set the variable that control the copy of custom meter definitions] ***\nThursday 26 March 2026  03:42:47 +0000 (0:00:00.861)       0:00:23.364 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Create default folder for custom meter definitions] *********\nThursday 26 March 2026  03:42:48 +0000 (0:00:00.685)       0:00:24.049 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying custom meter definitions to Ceilometer] *************\nThursday 26 March 2026  03:42:49 +0000 (0:00:01.070)       0:00:25.120 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Check if the folder [\"/opt/configuration/environments/kolla/files/overlays/ceilometer/pollsters.d\"] for dynamic pollsters definitions exist] ***\nThursday 26 March 2026  03:42:50 +0000 (0:00:00.871)       0:00:25.991 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceilometer : Set the variable that control the copy of dynamic pollsters definitions] ***\nThursday 26 March 2026  03:42:50 +0000 (0:00:00.712)       0:00:26.703 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Clean default folder for dynamic pollsters definitions] *****\nThursday 26 March 2026  03:42:51 +0000 (0:00:00.846)       0:00:27.550 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Create default folder for dynamic pollsters definitions] ****\nThursday 26 March 2026  03:42:52 +0000 (0:00:00.943)       0:00:28.494 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying dynamic pollsters definitions] **********************\nThursday 26 March 2026  03:42:53 +0000 (0:00:00.835)       0:00:29.329 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Check if custom polling.yaml exists] ************************\nThursday 26 March 2026  03:42:54 +0000 (0:00:00.625)       0:00:29.954 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [ceilometer : Copying over polling.yaml] **********************************\nThursday 26 March 2026  03:42:55 +0000 (0:00:01.632)       0:00:31.586 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Set ceilometer polling file's path] *************************\nThursday 26 March 2026  03:42:56 +0000 (0:00:00.814)       0:00:32.400 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Check custom gnocchi_resources.yaml exists] *****************\nThursday 26 March 2026  03:42:57 +0000 (0:00:00.823)       0:00:33.224 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [ceilometer : Copying over gnocchi_resources.yaml] ************************\nThursday 26 March 2026  03:42:58 +0000 (0:00:01.344)       0:00:34.569 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Set ceilometer gnocchi_resources file's path] ***************\nThursday 26 March 2026  03:42:59 +0000 (0:00:01.094)       0:00:35.664 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Check if policies shall be overwritten] *********************\nThursday 26 March 2026  03:43:00 +0000 (0:00:00.877)       0:00:36.542 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceilometer : Set ceilometer policy file] *********************************\nThursday 26 March 2026  03:43:00 +0000 (0:00:00.148)       0:00:36.690 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : include_tasks] **********************************************\nThursday 26 March 2026  03:43:01 +0000 (0:00:00.651)       0:00:37.341 ******** \n\u001b[0;36mincluded: /ansible/roles/ceilometer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : ceilometer | Copying over extra CA certificates] *****\nThursday 26 March 2026  03:43:02 +0000 (0:00:01.357)       0:00:38.699 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : ceilometer | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:43:05 +0000 (0:00:02.407)       0:00:41.106 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : ceilometer | Copying over backend internal TLS key] ***\nThursday 26 March 2026  03:43:06 +0000 (0:00:00.923)       0:00:42.029 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying over config.json files for services] ****************\nThursday 26 March 2026  03:43:07 +0000 (0:00:01.596)       0:00:43.626 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\nTASK [ceilometer : Copying over ceilometer.conf] *******************************\nThursday 26 March 2026  03:43:10 +0000 (0:00:02.796)       0:00:46.422 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\nTASK [ceilometer : Check custom event_definitions.yaml exists] *****************\nThursday 26 March 2026  03:43:15 +0000 (0:00:05.232)       0:00:51.654 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [ceilometer : Copying over event_definitions.yaml] ************************\nThursday 26 March 2026  03:43:17 +0000 (0:00:01.732)       0:00:53.386 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying over event_definitions.yaml for notification service] ***\nThursday 26 March 2026  03:43:18 +0000 (0:00:00.613)       0:00:54.000 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceilometer : Copying over event_pipeline.yaml] ***************************\nThursday 26 March 2026  03:43:19 +0000 (0:00:01.664)       0:00:55.665 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceilometer : Check custom pipeline.yaml exists] **************************\nThursday 26 March 2026  03:43:21 +0000 (0:00:01.535)       0:00:57.201 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [ceilometer : Copying over custom pipeline.yaml file] *********************\nThursday 26 March 2026  03:43:22 +0000 (0:00:01.620)       0:00:58.821 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\nTASK [ceilometer : Copying over pipeline.yaml file] ****************************\nThursday 26 March 2026  03:43:25 +0000 (0:00:02.674)       0:01:01.495 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying VMware vCenter CA file] *****************************\nThursday 26 March 2026  03:43:26 +0000 (0:00:00.881)       0:01:02.377 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Copying over existing policy file] **************************\nThursday 26 March 2026  03:43:27 +0000 (0:00:00.836)       0:01:03.214 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Check ceilometer containers] ********************************\nThursday 26 March 2026  03:43:28 +0000 (0:00:00.862)       0:01:04.076 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m\n\nTASK [ceilometer : include_tasks] **********************************************\nThursday 26 March 2026  03:43:29 +0000 (0:00:01.655)       0:01:05.731 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceilometer : Running Ceilometer bootstrap container] *********************\nThursday 26 March 2026  03:43:30 +0000 (0:00:00.653)       0:01:06.385 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:04.792)       0:01:11.178 ******** \n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:00.073)       0:01:11.251 ******** \n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:00.109)       0:01:11.361 ******** \n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:00.254)       0:01:11.615 ******** \n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:00.071)       0:01:11.686 ******** \n\nTASK [ceilometer : Flush handlers] *********************************************\nThursday 26 March 2026  03:43:35 +0000 (0:00:00.070)       0:01:11.757 ******** \n\nRUNNING HANDLER [ceilometer : Restart ceilometer-notification container] *******\nThursday 26 March 2026  03:43:36 +0000 (0:00:00.074)       0:01:11.831 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceilometer : Restart ceilometer-central container] ************\nThursday 26 March 2026  03:43:43 +0000 (0:00:07.594)       0:01:19.425 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceilometer : Restart ceilometer-compute container] ************\nThursday 26 March 2026  03:43:48 +0000 (0:00:04.951)       0:01:24.377 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:43:59 +0000 (0:00:11.229)       0:01:35.606 ******** \n=============================================================================== \nceilometer : Restart ceilometer-compute container ---------------------- 11.23s\nceilometer : Restart ceilometer-notification container ------------------ 7.59s\nceilometer : Copying over ceilometer.conf ------------------------------- 5.23s\nceilometer : Restart ceilometer-central container ----------------------- 4.95s\nceilometer : Running Ceilometer bootstrap container --------------------- 4.79s\nservice-ks-register : ceilometer | Granting user roles ------------------ 4.31s\nservice-ks-register : ceilometer | Creating users ----------------------- 4.00s\nservice-ks-register : ceilometer | Creating projects -------------------- 3.89s\nservice-ks-register : ceilometer | Creating roles ----------------------- 3.23s\nceilometer : Copying over config.json files for services ---------------- 2.80s\nceilometer : Copying over custom pipeline.yaml file --------------------- 2.67s\nservice-cert-copy : ceilometer | Copying over extra CA certificates ----- 2.41s\nceilometer : Check custom event_definitions.yaml exists ----------------- 1.73s\nceilometer : Check if the folder for custom meter definitions exist ----- 1.68s\nceilometer : Copying over event_definitions.yaml for notification service --- 1.66s\nceilometer : Check ceilometer containers -------------------------------- 1.66s\nceilometer : Check if custom polling.yaml exists ------------------------ 1.63s\nceilometer : Check custom pipeline.yaml exists -------------------------- 1.62s\nservice-cert-copy : ceilometer | Copying over backend internal TLS key --- 1.60s\nceilometer : Copying over event_pipeline.yaml --------------------------- 1.54s\n\u001b[32m2026-03-26 03:44:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dff533f4-b9a6-47d4-badc-74d13e9ba245 (aodh) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:44:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dff533f4-b9a6-47d4-badc-74d13e9ba245 (aodh) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:44:07 +0000 (0:00:00.286)       0:00:00.286 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:44:07 +0000 (0:00:00.309)       0:00:00.596 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_aodh_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_aodh_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_aodh_True)\u001b[0m\n\nPLAY [Apply role aodh] *********************************************************\n\nTASK [aodh : include_tasks] ****************************************************\nThursday 26 March 2026  03:44:08 +0000 (0:00:00.450)       0:00:01.046 ******** \n\u001b[0;36mincluded: /ansible/roles/aodh/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : aodh | Creating services] **************************\nThursday 26 March 2026  03:44:08 +0000 (0:00:00.622)       0:00:01.669 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=aodh (alarming))\u001b[0m\n\nTASK [service-ks-register : aodh | Creating endpoints] *************************\nThursday 26 March 2026  03:44:12 +0000 (0:00:03.760)       0:00:05.430 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> https://api-int.testbed.osism.xyz:8042 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> https://api.testbed.osism.xyz:8042 -> public)\u001b[0m\n\nTASK [service-ks-register : aodh | Creating projects] **************************\nThursday 26 March 2026  03:44:19 +0000 (0:00:06.645)       0:00:12.076 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : aodh | Creating users] *****************************\nThursday 26 March 2026  03:44:22 +0000 (0:00:03.452)       0:00:15.529 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> service)\u001b[0m\n\nTASK [service-ks-register : aodh | Creating roles] *****************************\nThursday 26 March 2026  03:44:26 +0000 (0:00:03.925)       0:00:19.454 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : aodh | Granting user roles] ************************\nThursday 26 March 2026  03:44:29 +0000 (0:00:03.429)       0:00:22.884 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> service -> admin)\u001b[0m\n\nTASK [aodh : Ensuring config directories exist] ********************************\nThursday 26 March 2026  03:44:33 +0000 (0:00:04.000)       0:00:26.885 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [aodh : Check if policies shall be overwritten] ***************************\nThursday 26 March 2026  03:44:36 +0000 (0:00:02.130)       0:00:29.016 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [aodh : Set aodh policy file] *********************************************\nThursday 26 March 2026  03:44:36 +0000 (0:00:00.132)       0:00:29.148 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [aodh : Copying over existing policy file] ********************************\nThursday 26 March 2026  03:44:36 +0000 (0:00:00.564)       0:00:29.712 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [aodh : include_tasks] ****************************************************\nThursday 26 March 2026  03:44:37 +0000 (0:00:00.677)       0:00:30.390 ******** \n\u001b[0;36mincluded: /ansible/roles/aodh/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : aodh | Copying over extra CA certificates] ***********\nThursday 26 March 2026  03:44:38 +0000 (0:00:00.801)       0:00:31.192 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : aodh | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:44:42 +0000 (0:00:04.462)       0:00:35.654 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : aodh | Copying over backend internal TLS key] ********\nThursday 26 March 2026  03:44:43 +0000 (0:00:00.684)       0:00:36.338 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [aodh : Copying over config.json files for services] **********************\nThursday 26 March 2026  03:44:44 +0000 (0:00:01.107)       0:00:37.445 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [aodh : Copying over aodh.conf] *******************************************\nThursday 26 March 2026  03:44:48 +0000 (0:00:04.195)       0:00:41.641 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [aodh : Copying over wsgi-aodh files for services] ************************\nThursday 26 March 2026  03:44:57 +0000 (0:00:08.707)       0:00:50.348 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [aodh : Check aodh containers] ********************************************\nThursday 26 March 2026  03:44:59 +0000 (0:00:01.790)       0:00:52.138 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [aodh : include_tasks] ****************************************************\nThursday 26 March 2026  03:45:02 +0000 (0:00:03.375)       0:00:55.514 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [aodh : Creating aodh database] *******************************************\nThursday 26 March 2026  03:45:02 +0000 (0:00:00.335)       0:00:55.849 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [aodh : Creating aodh database user and setting permissions] **************\nThursday 26 March 2026  03:45:05 +0000 (0:00:02.153)       0:00:58.003 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [aodh : Running aodh bootstrap container] *********************************\nThursday 26 March 2026  03:45:07 +0000 (0:00:02.269)       0:01:00.273 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [aodh : Flush handlers] ***************************************************\nThursday 26 March 2026  03:45:20 +0000 (0:00:13.524)       0:01:13.797 ******** \n\nTASK [aodh : Flush handlers] ***************************************************\nThursday 26 March 2026  03:45:20 +0000 (0:00:00.087)       0:01:13.884 ******** \n\nTASK [aodh : Flush handlers] ***************************************************\nThursday 26 March 2026  03:45:20 +0000 (0:00:00.072)       0:01:13.957 ******** \n\nRUNNING HANDLER [aodh : Restart aodh-api container] ****************************\nThursday 26 March 2026  03:45:21 +0000 (0:00:00.289)       0:01:14.246 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [aodh : Restart aodh-evaluator container] **********************\nThursday 26 March 2026  03:45:31 +0000 (0:00:10.640)       0:01:24.887 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [aodh : Restart aodh-listener container] ***********************\nThursday 26 March 2026  03:45:40 +0000 (0:00:08.230)       0:01:33.118 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [aodh : Restart aodh-notifier container] ***********************\nThursday 26 March 2026  03:45:50 +0000 (0:00:10.712)       0:01:43.831 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=17  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:46:01 +0000 (0:00:10.559)       0:01:54.390 ******** \n=============================================================================== \naodh : Running aodh bootstrap container -------------------------------- 13.52s\naodh : Restart aodh-listener container --------------------------------- 10.71s\naodh : Restart aodh-api container -------------------------------------- 10.64s\naodh : Restart aodh-notifier container --------------------------------- 10.56s\naodh : Copying over aodh.conf ------------------------------------------- 8.71s\naodh : Restart aodh-evaluator container --------------------------------- 8.23s\nservice-ks-register : aodh | Creating endpoints ------------------------- 6.65s\nservice-cert-copy : aodh | Copying over extra CA certificates ----------- 4.46s\naodh : Copying over config.json files for services ---------------------- 4.20s\nservice-ks-register : aodh | Granting user roles ------------------------ 4.00s\nservice-ks-register : aodh | Creating users ----------------------------- 3.93s\nservice-ks-register : aodh | Creating services -------------------------- 3.76s\nservice-ks-register : aodh | Creating projects -------------------------- 3.45s\nservice-ks-register : aodh | Creating roles ----------------------------- 3.43s\naodh : Check aodh containers -------------------------------------------- 3.38s\naodh : Creating aodh database user and setting permissions -------------- 2.27s\naodh : Creating aodh database ------------------------------------------- 2.15s\naodh : Ensuring config directories exist -------------------------------- 2.13s\naodh : Copying over wsgi-aodh files for services ------------------------ 1.79s\nservice-cert-copy : aodh | Copying over backend internal TLS key -------- 1.11s\n\u001b[32m2026-03-26 03:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac547b8e-a406-49a2-abee-9b9679420316 (kolla-ceph-rgw) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ac547b8e-a406-49a2-abee-9b9679420316 (kolla-ceph-rgw) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:46:08 +0000 (0:00:00.316)       0:00:00.316 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:46:09 +0000 (0:00:00.955)       0:00:01.272 ******** \n\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-rgw : include_tasks] ************************************************\nThursday 26 March 2026  03:46:10 +0000 (0:00:00.822)       0:00:02.094 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating services] **********************\nThursday 26 March 2026  03:46:12 +0000 (0:00:01.656)       0:00:03.751 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating endpoints] *********************\nThursday 26 March 2026  03:46:16 +0000 (0:00:03.922)       0:00:07.673 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating projects] **********************\nThursday 26 March 2026  03:46:22 +0000 (0:00:06.506)       0:00:14.180 ******** \n\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating users] *************************\nThursday 26 March 2026  03:46:25 +0000 (0:00:03.325)       0:00:17.505 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating roles] *************************\nThursday 26 March 2026  03:46:29 +0000 (0:00:04.015)       0:00:21.521 ******** \n\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Granting user roles] ********************\nThursday 26 March 2026  03:46:36 +0000 (0:00:06.487)       0:00:28.008 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:46:41 +0000 (0:00:05.109)       0:00:33.118 ******** \n=============================================================================== \nservice-ks-register : ceph-rgw | Creating endpoints --------------------- 6.51s\nservice-ks-register : ceph-rgw | Creating roles ------------------------- 6.49s\nservice-ks-register : ceph-rgw | Granting user roles -------------------- 5.11s\nservice-ks-register : ceph-rgw | Creating users ------------------------- 4.02s\nservice-ks-register : ceph-rgw | Creating services ---------------------- 3.92s\nservice-ks-register : ceph-rgw | Creating projects ---------------------- 3.33s\nceph-rgw : include_tasks ------------------------------------------------ 1.66s\nGroup hosts based on Kolla action --------------------------------------- 0.96s\nGroup hosts based on enabled services ----------------------------------- 0.82s\n\u001b[32m2026-03-26 03:46:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cab4e368-13de-4ef3-b76d-5a075b924f38 (gnocchi) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:46:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cab4e368-13de-4ef3-b76d-5a075b924f38 (gnocchi) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:46:49 +0000 (0:00:00.288)       0:00:00.288 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:46:49 +0000 (0:00:00.336)       0:00:00.625 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_gnocchi_False)\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_gnocchi_True\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_gnocchi_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_gnocchi_False)\u001b[0m\n\nPLAY [Apply role gnocchi] ******************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:46:49 +0000 (0:00:00.415)       0:00:01.040 ******** \n=============================================================================== \nGroup hosts based on enabled services ----------------------------------- 0.42s\nGroup hosts based on Kolla action --------------------------------------- 0.34s\n\u001b[32m2026-03-26 03:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4dad45d8-ad3c-4a11-bb8f-ae77d3a40d56 (manila) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4dad45d8-ad3c-4a11-bb8f-ae77d3a40d56 (manila) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:46:57 +0000 (0:00:00.291)       0:00:00.291 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:46:57 +0000 (0:00:00.320)       0:00:00.611 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_manila_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_manila_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_manila_True)\u001b[0m\n\nPLAY [Apply role manila] *******************************************************\n\nTASK [manila : include_tasks] **************************************************\nThursday 26 March 2026  03:46:57 +0000 (0:00:00.466)       0:00:01.078 ******** \n\u001b[0;36mincluded: /ansible/roles/manila/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [manila : include_tasks] **************************************************\nThursday 26 March 2026  03:46:58 +0000 (0:00:00.575)       0:00:01.654 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-ks-register : manila | Creating services] ************************\nThursday 26 March 2026  03:46:59 +0000 (0:00:00.494)       0:00:02.148 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=manila (share))\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 (sharev2))\u001b[0m\n\nTASK [service-ks-register : manila | Creating endpoints] ***********************\nThursday 26 March 2026  03:47:05 +0000 (0:00:06.592)       0:00:08.741 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> https://api-int.testbed.osism.xyz:8786/v1/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> https://api.testbed.osism.xyz:8786/v1/%(tenant_id)s -> public)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 -> https://api-int.testbed.osism.xyz:8786/v2 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 -> https://api.testbed.osism.xyz:8786/v2 -> public)\u001b[0m\n\nTASK [service-ks-register : manila | Creating projects] ************************\nThursday 26 March 2026  03:47:18 +0000 (0:00:13.260)       0:00:22.002 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : manila | Creating users] ***************************\nThursday 26 March 2026  03:47:22 +0000 (0:00:03.416)       0:00:25.418 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> service)\u001b[0m\n\nTASK [service-ks-register : manila | Creating roles] ***************************\nThursday 26 March 2026  03:47:26 +0000 (0:00:03.981)       0:00:29.400 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : manila | Granting user roles] **********************\nThursday 26 March 2026  03:47:29 +0000 (0:00:03.226)       0:00:32.626 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> service -> admin)\u001b[0m\n\nTASK [manila : Ensuring config directories exist] ******************************\nThursday 26 March 2026  03:47:33 +0000 (0:00:04.011)       0:00:36.638 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [manila : include_tasks] **************************************************\nThursday 26 March 2026  03:47:35 +0000 (0:00:02.411)       0:00:39.049 ******** \n\u001b[0;36mincluded: /ansible/roles/manila/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [manila : Ensuring manila service ceph config subdir exists] **************\nThursday 26 March 2026  03:47:36 +0000 (0:00:00.633)       0:00:39.683 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [manila : Copy over multiple ceph configs for Manila] *********************\nThursday 26 March 2026  03:47:37 +0000 (0:00:01.053)       0:00:40.736 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\nTASK [manila : Copy over ceph Manila keyrings] *********************************\nThursday 26 March 2026  03:47:39 +0000 (0:00:01.877)       0:00:42.613 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m\n\nTASK [manila : Ensuring config directory has correct owner and permission] *****\nThursday 26 March 2026  03:47:40 +0000 (0:00:01.295)       0:00:43.909 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=manila-share)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=manila-share)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=manila-share)\u001b[0m\n\nTASK [manila : Check if policies shall be overwritten] *************************\nThursday 26 March 2026  03:47:41 +0000 (0:00:00.846)       0:00:44.755 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [manila : Set manila policy file] *****************************************\nThursday 26 March 2026  03:47:41 +0000 (0:00:00.155)       0:00:44.910 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [manila : include_tasks] **************************************************\nThursday 26 March 2026  03:47:42 +0000 (0:00:00.684)       0:00:45.595 ******** \n\u001b[0;36mincluded: /ansible/roles/manila/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : manila | Copying over extra CA certificates] *********\nThursday 26 March 2026  03:47:43 +0000 (0:00:00.715)       0:00:46.310 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : manila | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:47:47 +0000 (0:00:04.172)       0:00:50.483 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : manila | Copying over backend internal TLS key] ******\nThursday 26 March 2026  03:47:48 +0000 (0:00:00.901)       0:00:51.384 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [manila : Copying over config.json files for services] ********************\nThursday 26 March 2026  03:47:49 +0000 (0:00:00.911)       0:00:52.296 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [manila : Copying over manila.conf] ***************************************\nThursday 26 March 2026  03:47:53 +0000 (0:00:04.662)       0:00:56.958 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [manila : Copying over manila-share.conf] *********************************\nThursday 26 March 2026  03:48:00 +0000 (0:00:07.104)       0:01:04.063 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=manila-share)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=manila-share)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=manila-share)\u001b[0m\n\nTASK [manila : Copying over existing policy file] ******************************\nThursday 26 March 2026  03:48:04 +0000 (0:00:03.914)       0:01:07.977 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [manila : Check manila containers] ****************************************\nThursday 26 March 2026  03:48:05 +0000 (0:00:00.767)       0:01:08.745 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [manila : Creating Manila database] ***************************************\nThursday 26 March 2026  03:48:09 +0000 (0:00:03.570)       0:01:12.315 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [manila : Creating Manila database user and setting permissions] **********\nThursday 26 March 2026  03:48:11 +0000 (0:00:02.477)       0:01:14.793 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [manila : Running Manila bootstrap container] *****************************\nThursday 26 March 2026  03:48:14 +0000 (0:00:02.303)       0:01:17.096 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [manila : Flush handlers] *************************************************\nThursday 26 March 2026  03:48:51 +0000 (0:00:37.486)       0:01:54.583 ******** \n\nTASK [manila : Flush handlers] *************************************************\nThursday 26 March 2026  03:48:51 +0000 (0:00:00.076)       0:01:54.660 ******** \n\nTASK [manila : Flush handlers] *************************************************\nThursday 26 March 2026  03:48:51 +0000 (0:00:00.074)       0:01:54.734 ******** \n\nRUNNING HANDLER [manila : Restart manila-api container] ************************\nThursday 26 March 2026  03:48:51 +0000 (0:00:00.077)       0:01:54.811 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [manila : Restart manila-data container] ***********************\nThursday 26 March 2026  03:49:07 +0000 (0:00:15.489)       0:02:10.300 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [manila : Restart manila-scheduler container] ******************\nThursday 26 March 2026  03:49:17 +0000 (0:00:10.691)       0:02:20.992 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [manila : Restart manila-share container] **********************\nThursday 26 March 2026  03:49:28 +0000 (0:00:10.408)       0:02:31.400 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=28  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:49:47 +0000 (0:00:19.047)       0:02:50.448 ******** \n=============================================================================== \nmanila : Running Manila bootstrap container ---------------------------- 37.49s\nmanila : Restart manila-share container -------------------------------- 19.05s\nmanila : Restart manila-api container ---------------------------------- 15.49s\nservice-ks-register : manila | Creating endpoints ---------------------- 13.26s\nmanila : Restart manila-data container --------------------------------- 10.69s\nmanila : Restart manila-scheduler container ---------------------------- 10.41s\nmanila : Copying over manila.conf --------------------------------------- 7.10s\nservice-ks-register : manila | Creating services ------------------------ 6.59s\nmanila : Copying over config.json files for services -------------------- 4.66s\nservice-cert-copy : manila | Copying over extra CA certificates --------- 4.17s\nservice-ks-register : manila | Granting user roles ---------------------- 4.01s\nservice-ks-register : manila | Creating users --------------------------- 3.98s\nmanila : Copying over manila-share.conf --------------------------------- 3.91s\nmanila : Check manila containers ---------------------------------------- 3.57s\nservice-ks-register : manila | Creating projects ------------------------ 3.42s\nservice-ks-register : manila | Creating roles --------------------------- 3.23s\nmanila : Creating Manila database --------------------------------------- 2.48s\nmanila : Ensuring config directories exist ------------------------------ 2.41s\nmanila : Creating Manila database user and setting permissions ---------- 2.30s\nmanila : Copy over multiple ceph configs for Manila --------------------- 1.88s\n+ sh -c /opt/configuration/scripts/deploy/400-monitoring.sh\n\u001b[32m2026-03-26 03:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0cae00d9-e756-4c7c-ac2c-a238d5320de9 (netdata) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0cae00d9-e756-4c7c-ac2c-a238d5320de9 (netdata) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:50:05 +0000 (0:00:00.303)       0:00:00.303 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m\n\nPLAY [Apply role netdata] ******************************************************\n\nTASK [osism.services.netdata : Include distribution specific install tasks] ****\nThursday 26 March 2026  03:50:06 +0000 (0:00:01.086)       0:00:01.389 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Remove old architecture-dependent repository] ***\nThursday 26 March 2026  03:50:08 +0000 (0:00:01.594)       0:00:02.984 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Install apt-transport-https package] ************\nThursday 26 March 2026  03:50:10 +0000 (0:00:02.156)       0:00:05.141 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Add repository gpg key] *************************\nThursday 26 March 2026  03:50:12 +0000 (0:00:02.351)       0:00:07.492 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add repository] *********************************\nThursday 26 March 2026  03:50:14 +0000 (0:00:01.655)       0:00:09.147 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.netdata : Install package netdata] ************************\nThursday 26 March 2026  03:50:30 +0000 (0:00:16.080)       0:00:25.228 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.netdata : Include config tasks] ***************************\nThursday 26 March 2026  03:51:12 +0000 (0:00:42.183)       0:01:07.411 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Copy configuration files] ***********************\nThursday 26 March 2026  03:51:14 +0000 (0:00:01.653)       0:01:09.064 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m\n\nTASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***\nThursday 26 March 2026  03:51:17 +0000 (0:00:03.435)       0:01:12.500 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Opt out from anonymous statistics] **************\nThursday 26 March 2026  03:51:19 +0000 (0:00:01.312)       0:01:13.812 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add netdata user to docker group] ***************\nThursday 26 March 2026  03:51:20 +0000 (0:00:01.334)       0:01:15.147 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Manage service netdata] *************************\nThursday 26 March 2026  03:51:21 +0000 (0:00:01.269)       0:01:16.417 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Include host type specific tasks] ***************\nThursday 26 March 2026  03:51:23 +0000 (0:00:01.788)       0:01:18.205 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********\nThursday 26 March 2026  03:51:25 +0000 (0:00:01.523)       0:01:19.729 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.netdata : Restart service netdata] *************\nThursday 26 March 2026  03:51:28 +0000 (0:00:03.358)       0:01:23.087 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:51:39 +0000 (0:00:11.180)       0:01:34.268 ******** \n=============================================================================== \nosism.services.netdata : Install package netdata ----------------------- 42.18s\nosism.services.netdata : Add repository -------------------------------- 16.08s\nosism.services.netdata : Restart service netdata ----------------------- 11.18s\nosism.services.netdata : Copy configuration files ----------------------- 3.44s\nosism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 3.36s\nosism.services.netdata : Install apt-transport-https package ------------ 2.35s\nosism.services.netdata : Remove old architecture-dependent repository --- 2.16s\nosism.services.netdata : Manage service netdata ------------------------- 1.79s\nosism.services.netdata : Add repository gpg key ------------------------- 1.66s\nosism.services.netdata : Include config tasks --------------------------- 1.65s\nosism.services.netdata : Include distribution specific install tasks ---- 1.59s\nosism.services.netdata : Include host type specific tasks --------------- 1.52s\nosism.services.netdata : Opt out from anonymous statistics -------------- 1.33s\nosism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.31s\nosism.services.netdata : Add netdata user to docker group --------------- 1.27s\nGroup hosts based on enabled services ----------------------------------- 1.09s\n\u001b[32m2026-03-26 03:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 841b3c3f-80c9-4ffe-85c0-5ab312ce4bb3 (prometheus) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 841b3c3f-80c9-4ffe-85c0-5ab312ce4bb3 (prometheus) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:51:50 +0000 (0:00:00.311)       0:00:00.311 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:51:52 +0000 (0:00:01.016)       0:00:01.328 ******** \n\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m\n\nPLAY [Apply role prometheus] ***************************************************\n\nTASK [prometheus : include_tasks] **********************************************\nThursday 26 March 2026  03:51:52 +0000 (0:00:00.954)       0:00:02.282 ******** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [prometheus : Ensuring config directories exist] **************************\nThursday 26 March 2026  03:51:54 +0000 (0:00:01.436)       0:00:03.718 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : include_tasks] **********************************************\nThursday 26 March 2026  03:51:57 +0000 (0:00:03.021)       0:00:06.740 ******** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****\nThursday 26 March 2026  03:51:59 +0000 (0:00:01.776)       0:00:08.516 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:52:04 +0000 (0:00:05.614)       0:00:14.130 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***\nThursday 26 March 2026  03:52:06 +0000 (0:00:02.178)       0:00:16.309 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over config.json files] *****************************\nThursday 26 March 2026  03:52:09 +0000 (0:00:02.365)       0:00:18.675 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Find custom prometheus alert rules files] *******************\nThursday 26 March 2026  03:52:16 +0000 (0:00:06.928)       0:00:25.603 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom prometheus alert rules files] ***********\nThursday 26 March 2026  03:52:17 +0000 (0:00:00.970)       0:00:26.573 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [prometheus : Find prometheus common config overrides] ********************\nThursday 26 March 2026  03:52:43 +0000 (0:00:25.818)       0:00:52.392 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Find prometheus host config overrides] **********************\nThursday 26 March 2026  03:52:43 +0000 (0:00:00.819)       0:00:53.211 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over prometheus config file] ************************\nThursday 26 March 2026  03:52:45 +0000 (0:00:01.919)       0:00:55.131 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus web config file] ********************\nThursday 26 March 2026  03:53:03 +0000 (0:00:17.839)       0:01:12.971 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus alertmanager config file] ***********\nThursday 26 March 2026  03:53:06 +0000 (0:00:02.942)       0:01:15.913 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m\n\nTASK [prometheus : Find custom Alertmanager alert notification templates] ******\nThursday 26 March 2026  03:53:08 +0000 (0:00:02.029)       0:01:17.943 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom Alertmanager alert notification templates] ***\nThursday 26 March 2026  03:53:09 +0000 (0:00:00.839)       0:01:18.782 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************\nThursday 26 March 2026  03:53:10 +0000 (0:00:00.868)       0:01:19.650 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [prometheus : Copying cloud config file for openstack exporter] ***********\nThursday 26 March 2026  03:53:12 +0000 (0:00:02.389)       0:01:22.040 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying config file for blackbox exporter] ******************\nThursday 26 March 2026  03:53:14 +0000 (0:00:01.614)       0:01:23.654 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m\n\nTASK [prometheus : Find extra prometheus server config files] ******************\nThursday 26 March 2026  03:53:16 +0000 (0:00:01.677)       0:01:25.332 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m\n\u001b[1;35mnot a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Create subdirectories for extra config files] ***************\nThursday 26 March 2026  03:53:17 +0000 (0:00:01.186)       0:01:26.518 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Template extra prometheus server config files] **************\nThursday 26 March 2026  03:53:18 +0000 (0:00:01.015)       0:01:27.534 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Check prometheus containers] ********************************\nThursday 26 March 2026  03:53:19 +0000 (0:00:00.943)       0:01:28.478 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Creating prometheus database user and setting permissions] ***\nThursday 26 March 2026  03:53:23 +0000 (0:00:04.294)       0:01:32.772 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:24 +0000 (0:00:01.479)       0:01:34.252 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.082)       0:01:34.335 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.077)       0:01:34.412 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.074)       0:01:34.486 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.080)       0:01:34.567 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.070)       0:01:34.637 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.082)       0:01:34.719 ******** \n\nRUNNING HANDLER [prometheus : Restart prometheus-server container] *************\nThursday 26 March 2026  03:53:25 +0000 (0:00:00.097)       0:01:34.817 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******\nThursday 26 March 2026  03:53:47 +0000 (0:00:21.534)       0:01:56.352 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****\nThursday 26 March 2026  03:54:00 +0000 (0:00:13.813)       0:02:10.165 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***\nThursday 26 March 2026  03:54:10 +0000 (0:00:10.085)       0:02:20.251 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********\nThursday 26 March 2026  03:54:21 +0000 (0:00:10.498)       0:02:30.749 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******\nThursday 26 March 2026  03:54:36 +0000 (0:00:14.650)       0:02:45.400 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***\nThursday 26 March 2026  03:54:45 +0000 (0:00:08.962)       0:02:54.362 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***\nThursday 26 March 2026  03:54:50 +0000 (0:00:05.845)       0:03:00.208 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***\nThursday 26 March 2026  03:54:56 +0000 (0:00:05.838)       0:03:06.047 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:55:07 +0000 (0:00:10.524)       0:03:16.572 ******** \n=============================================================================== \nprometheus : Copying over custom prometheus alert rules files ---------- 25.82s\nprometheus : Restart prometheus-server container ----------------------- 21.53s\nprometheus : Copying over prometheus config file ----------------------- 17.84s\nprometheus : Restart prometheus-cadvisor container --------------------- 14.65s\nprometheus : Restart prometheus-node-exporter container ---------------- 13.81s\nprometheus : Restart prometheus-libvirt-exporter container ------------- 10.52s\nprometheus : Restart prometheus-memcached-exporter container ----------- 10.50s\nprometheus : Restart prometheus-mysqld-exporter container -------------- 10.09s\nprometheus : Restart prometheus-alertmanager container ------------------ 8.96s\nprometheus : Copying over config.json files ----------------------------- 6.93s\nprometheus : Restart prometheus-elasticsearch-exporter container -------- 5.85s\nprometheus : Restart prometheus-blackbox-exporter container ------------- 5.84s\nservice-cert-copy : prometheus | Copying over extra CA certificates ----- 5.61s\nprometheus : Check prometheus containers -------------------------------- 4.29s\nprometheus : Ensuring config directories exist -------------------------- 3.02s\nprometheus : Copying over prometheus web config file -------------------- 2.94s\nprometheus : Copying over my.cnf for mysqld_exporter -------------------- 2.39s\nservice-cert-copy : prometheus | Copying over backend internal TLS key --- 2.37s\nservice-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.18s\nprometheus : Copying over prometheus alertmanager config file ----------- 2.03s\n\u001b[32m2026-03-26 03:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 00c4bf35-4c60-4b40-be13-05f79dea06c0 (grafana) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 00c4bf35-4c60-4b40-be13-05f79dea06c0 (grafana) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:55:14 +0000 (0:00:00.255)       0:00:00.255 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:55:14 +0000 (0:00:00.375)       0:00:00.630 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m\n\nPLAY [Apply role grafana] ******************************************************\n\nTASK [grafana : include_tasks] *************************************************\nThursday 26 March 2026  03:55:14 +0000 (0:00:00.484)       0:00:01.114 ******** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [grafana : Ensuring config directories exist] *****************************\nThursday 26 March 2026  03:55:15 +0000 (0:00:00.605)       0:00:01.720 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Check if extra configuration file exists] **********************\nThursday 26 March 2026  03:55:16 +0000 (0:00:00.923)       0:00:02.643 ******** \n\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m\n\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : include_tasks] *************************************************\nThursday 26 March 2026  03:55:17 +0000 (0:00:00.910)       0:00:03.554 ******** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over extra CA certificates] ********\nThursday 26 March 2026  03:55:18 +0000 (0:00:00.642)       0:00:04.196 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:55:19 +0000 (0:00:01.379)       0:00:05.575 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****\nThursday 26 March 2026  03:55:20 +0000 (0:00:00.638)       0:00:06.214 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over config.json files] ********************************\nThursday 26 March 2026  03:55:20 +0000 (0:00:00.666)       0:00:06.881 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over grafana.ini] **************************************\nThursday 26 March 2026  03:55:22 +0000 (0:00:01.369)       0:00:08.251 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over extra configuration file] *************************\nThursday 26 March 2026  03:55:23 +0000 (0:00:01.769)       0:00:10.021 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Configuring Prometheus as data source for Grafana] *************\nThursday 26 March 2026  03:55:24 +0000 (0:00:00.390)       0:00:10.411 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\nTASK [grafana : Configuring dashboards provisioning] ***************************\nThursday 26 March 2026  03:55:25 +0000 (0:00:01.360)       0:00:11.771 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\nTASK [grafana : Find custom grafana dashboards] ********************************\nThursday 26 March 2026  03:55:27 +0000 (0:00:01.892)       0:00:13.664 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : Find templated grafana dashboards] *****************************\nThursday 26 March 2026  03:55:28 +0000 (0:00:00.809)       0:00:14.474 ******** \n\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m\n\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Prune templated Grafana dashboards] ****************************\nThursday 26 March 2026  03:55:29 +0000 (0:00:00.739)       0:00:15.213 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over custom dashboards] ********************************\nThursday 26 March 2026  03:55:29 +0000 (0:00:00.378)       0:00:15.593 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\nTASK [grafana : Check grafana containers] **************************************\nThursday 26 March 2026  03:56:08 +0000 (0:00:38.713)       0:00:54.306 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Creating grafana database] *************************************\nThursday 26 March 2026  03:56:09 +0000 (0:00:01.098)       0:00:55.404 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Creating grafana database user and setting permissions] ********\nThursday 26 March 2026  03:56:11 +0000 (0:00:02.371)       0:00:57.776 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Flush handlers] ************************************************\nThursday 26 March 2026  03:56:14 +0000 (0:00:02.393)       0:01:00.169 ******** \n\nTASK [grafana : Flush handlers] ************************************************\nThursday 26 March 2026  03:56:14 +0000 (0:00:00.094)       0:01:00.264 ******** \n\nTASK [grafana : Flush handlers] ************************************************\nThursday 26 March 2026  03:56:14 +0000 (0:00:00.106)       0:01:00.370 ******** \n\nRUNNING HANDLER [grafana : Restart first grafana container] ********************\nThursday 26 March 2026  03:56:14 +0000 (0:00:00.085)       0:01:00.456 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********\nThursday 26 March 2026  03:56:16 +0000 (0:00:02.293)       0:01:02.749 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (9 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Restart remaining grafana containers] ***************\nThursday 26 March 2026  03:57:07 +0000 (0:00:51.148)       0:01:53.898 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Wait for grafana application ready] ****************************\nThursday 26 March 2026  03:57:41 +0000 (0:00:33.842)       0:02:27.740 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Remove old grafana docker volume] ******************************\nThursday 26 March 2026  03:57:44 +0000 (0:00:02.374)       0:02:30.115 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Enable grafana datasources] ************************************\nThursday 26 March 2026  03:57:44 +0000 (0:00:00.356)       0:02:30.471 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m\n\nTASK [grafana : Disable Getting Started panel] *********************************\nThursday 26 March 2026  03:57:47 +0000 (0:00:02.724)       0:02:33.195 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:57:47 +0000 (0:00:00.328)       0:02:33.523 ******** \n=============================================================================== \ngrafana : Waiting for grafana to start on first node ------------------- 51.15s\ngrafana : Copying over custom dashboards ------------------------------- 38.71s\ngrafana : Restart remaining grafana containers ------------------------- 33.84s\ngrafana : Enable grafana datasources ------------------------------------ 2.72s\ngrafana : Creating grafana database user and setting permissions -------- 2.39s\ngrafana : Wait for grafana application ready ---------------------------- 2.37s\ngrafana : Creating grafana database ------------------------------------- 2.37s\ngrafana : Restart first grafana container ------------------------------- 2.29s\ngrafana : Configuring dashboards provisioning --------------------------- 1.89s\ngrafana : Copying over grafana.ini -------------------------------------- 1.77s\nservice-cert-copy : grafana | Copying over extra CA certificates -------- 1.38s\ngrafana : Copying over config.json files -------------------------------- 1.37s\ngrafana : Configuring Prometheus as data source for Grafana ------------- 1.36s\ngrafana : Check grafana containers -------------------------------------- 1.10s\ngrafana : Ensuring config directories exist ----------------------------- 0.92s\ngrafana : Check if extra configuration file exists ---------------------- 0.91s\ngrafana : Find custom grafana dashboards -------------------------------- 0.81s\ngrafana : Find templated grafana dashboards ----------------------------- 0.74s\nservice-cert-copy : grafana | Copying over backend internal TLS key ----- 0.67s\ngrafana : include_tasks ------------------------------------------------- 0.64s\n+ sh -c /opt/configuration/scripts/deploy/510-clusterapi.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n++ semver 9.5.0 8.0.0\n+ [[ 1 -ge 0 ]]\n+ osism apply clusterapi\n\u001b[32m2026-03-26 03:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8d0762b2-3f92-45d4-a9d6-72554dc03cb3 (clusterapi) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8d0762b2-3f92-45d4-a9d6-72554dc03cb3 (clusterapi) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply cert_manager role] *************************************************\n\nTASK [Include cert_manager role] ***********************************************\nThursday 26 March 2026  03:57:55 +0000 (0:00:00.208)       0:00:00.208 ******** \n\u001b[0;36mincluded: cert_manager for testbed-manager\u001b[0m\n\nTASK [cert_manager : Deploy cert-manager crds] *********************************\nThursday 26 March 2026  03:57:56 +0000 (0:00:00.262)       0:00:00.470 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [cert_manager : Deploy cert-manager] **************************************\nThursday 26 March 2026  03:58:01 +0000 (0:00:05.815)       0:00:06.285 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Initialize or upgrade the CAPI management cluster] ***********************\n\nTASK [Get capi-system namespace phase] *****************************************\nThursday 26 March 2026  03:58:26 +0000 (0:00:24.656)       0:00:30.942 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Set capi-system-phase fact] **********************************************\nThursday 26 March 2026  03:58:27 +0000 (0:00:01.152)       0:00:32.094 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Initialize the CAPI management cluster] **********************************\nThursday 26 March 2026  03:58:27 +0000 (0:00:00.170)       0:00:32.265 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Upgrade the CAPI management cluster] *************************************\nThursday 26 March 2026  03:58:44 +0000 (0:00:16.373)       0:00:48.639 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [Install openstack-resource-controller] ***********************************\nThursday 26 March 2026  03:58:44 +0000 (0:00:00.169)       0:00:48.808 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  03:58:46 +0000 (0:00:02.321)       0:00:51.130 ******** \n=============================================================================== \ncert_manager : Deploy cert-manager ------------------------------------- 24.66s\nInitialize the CAPI management cluster --------------------------------- 16.37s\ncert_manager : Deploy cert-manager crds --------------------------------- 5.82s\nInstall openstack-resource-controller ----------------------------------- 2.32s\nGet capi-system namespace phase ----------------------------------------- 1.15s\nInclude cert_manager role ----------------------------------------------- 0.26s\nSet capi-system-phase fact ---------------------------------------------- 0.17s\nUpgrade the CAPI management cluster ------------------------------------- 0.17s\n+ osism apply magnum\n\u001b[32m2026-03-26 03:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 74b26c9e-e173-4941-b638-fff6166fc4d3 (magnum) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 03:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 74b26c9e-e173-4941-b638-fff6166fc4d3 (magnum) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  03:58:54 +0000 (0:00:00.301)       0:00:00.301 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  03:58:54 +0000 (0:00:00.336)       0:00:00.638 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m\n\nPLAY [Apply role magnum] *******************************************************\n\nTASK [magnum : include_tasks] **************************************************\nThursday 26 March 2026  03:58:55 +0000 (0:00:00.529)       0:00:01.167 ******** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : magnum | Creating services] ************************\nThursday 26 March 2026  03:58:55 +0000 (0:00:00.669)       0:00:01.837 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m\n\nTASK [service-ks-register : magnum | Creating endpoints] ***********************\nThursday 26 March 2026  03:58:59 +0000 (0:00:03.626)       0:00:05.464 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating projects] ************************\nThursday 26 March 2026  03:59:05 +0000 (0:00:06.472)       0:00:11.936 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating users] ***************************\nThursday 26 March 2026  03:59:09 +0000 (0:00:03.545)       0:00:15.482 ******** \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating roles] ***************************\nThursday 26 March 2026  03:59:13 +0000 (0:00:03.939)       0:00:19.422 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : magnum | Granting user roles] **********************\nThursday 26 March 2026  03:59:16 +0000 (0:00:03.416)       0:00:22.838 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m\n\nTASK [magnum : Creating Magnum trustee domain] *********************************\nThursday 26 March 2026  03:59:21 +0000 (0:00:04.371)       0:00:27.210 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user] ***********************************\nThursday 26 March 2026  03:59:24 +0000 (0:00:03.622)       0:00:30.832 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user role] ******************************\nThursday 26 March 2026  03:59:28 +0000 (0:00:03.854)       0:00:34.687 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Ensuring config directories exist] ******************************\nThursday 26 March 2026  03:59:32 +0000 (0:00:03.448)       0:00:38.135 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Check if policies shall be overwritten] *************************\nThursday 26 March 2026  03:59:33 +0000 (0:00:01.669)       0:00:39.805 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Set magnum policy file] *****************************************\nThursday 26 March 2026  03:59:33 +0000 (0:00:00.189)       0:00:39.995 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check if kubeconfig file is supplied] ***************************\nThursday 26 March 2026  03:59:34 +0000 (0:00:00.341)       0:00:40.336 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [magnum : Copying over kubeconfig file] ***********************************\nThursday 26 March 2026  03:59:35 +0000 (0:00:00.911)       0:00:41.248 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Set magnum kubeconfig file's path] ******************************\nThursday 26 March 2026  03:59:37 +0000 (0:00:02.571)       0:00:43.819 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nThursday 26 March 2026  03:59:38 +0000 (0:00:00.548)       0:00:44.367 ******** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over extra CA certificates] *********\nThursday 26 March 2026  03:59:38 +0000 (0:00:00.582)       0:00:44.950 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  03:59:41 +0000 (0:00:02.655)       0:00:47.605 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******\nThursday 26 March 2026  03:59:42 +0000 (0:00:01.094)       0:00:48.699 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Copying over config.json files for services] ********************\nThursday 26 March 2026  03:59:43 +0000 (0:00:00.934)       0:00:49.633 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over magnum.conf] ***************************************\nThursday 26 March 2026  03:59:46 +0000 (0:00:02.739)       0:00:52.373 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over existing policy file] ******************************\nThursday 26 March 2026  03:59:51 +0000 (0:00:05.740)       0:00:58.113 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check magnum containers] ****************************************\nThursday 26 March 2026  03:59:52 +0000 (0:00:00.797)       0:00:58.910 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nThursday 26 March 2026  03:59:54 +0000 (0:00:01.979)       0:01:00.889 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Creating Magnum database] ***************************************\nThursday 26 March 2026  03:59:55 +0000 (0:00:00.754)       0:01:01.644 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database user and setting permissions] **********\nThursday 26 March 2026  03:59:57 +0000 (0:00:02.176)       0:01:03.820 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Running Magnum bootstrap container] *****************************\nThursday 26 March 2026  04:00:00 +0000 (0:00:02.444)       0:01:06.265 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Flush handlers] *************************************************\nThursday 26 March 2026  04:00:17 +0000 (0:00:17.263)       0:01:23.529 ******** \n\nTASK [magnum : Flush handlers] *************************************************\nThursday 26 March 2026  04:00:17 +0000 (0:00:00.089)       0:01:23.618 ******** \n\nTASK [magnum : Flush handlers] *************************************************\nThursday 26 March 2026  04:00:17 +0000 (0:00:00.076)       0:01:23.695 ******** \n\nRUNNING HANDLER [magnum : Restart magnum-api container] ************************\nThursday 26 March 2026  04:00:17 +0000 (0:00:00.075)       0:01:23.771 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [magnum : Restart magnum-conductor container] ******************\nThursday 26 March 2026  04:00:38 +0000 (0:00:20.763)       0:01:44.534 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:00:50 +0000 (0:00:11.676)       0:01:56.211 ******** \n=============================================================================== \nmagnum : Restart magnum-api container ---------------------------------- 20.76s\nmagnum : Running Magnum bootstrap container ---------------------------- 17.26s\nmagnum : Restart magnum-conductor container ---------------------------- 11.68s\nservice-ks-register : magnum | Creating endpoints ----------------------- 6.47s\nmagnum : Copying over magnum.conf --------------------------------------- 5.74s\nservice-ks-register : magnum | Granting user roles ---------------------- 4.37s\nservice-ks-register : magnum | Creating users --------------------------- 3.94s\nmagnum : Creating Magnum trustee user ----------------------------------- 3.85s\nservice-ks-register : magnum | Creating services ------------------------ 3.63s\nmagnum : Creating Magnum trustee domain --------------------------------- 3.62s\nservice-ks-register : magnum | Creating projects ------------------------ 3.55s\nmagnum : Creating Magnum trustee user role ------------------------------ 3.45s\nservice-ks-register : magnum | Creating roles --------------------------- 3.42s\nmagnum : Copying over config.json files for services -------------------- 2.74s\nservice-cert-copy : magnum | Copying over extra CA certificates --------- 2.66s\nmagnum : Copying over kubeconfig file ----------------------------------- 2.57s\nmagnum : Creating Magnum database user and setting permissions ---------- 2.44s\nmagnum : Creating Magnum database --------------------------------------- 2.18s\nmagnum : Check magnum containers ---------------------------------------- 1.98s\nmagnum : Ensuring config directories exist ------------------------------ 1.67s",
                            "stdout_lines": [
                                "",
                                "# DEPLOY SERVICES",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY SERVICES'",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sh -c /opt/configuration/scripts/pull-images.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "",
                                "# PULL IMAGES",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "+ echo",
                                "",
                                "++ semver 9.5.0 7.0.0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-03-26 02:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-03-26 02:15:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06102fcd-70e1-4004-8e7e-1d35d1447f80 (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:15:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06102fcd-70e1-4004-8e7e-1d35d1447f80 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/deploy/001-helpers.sh",
                                "\u001b[32m2026-03-26 02:15:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0e8cb748-e270-4d9d-9e6d-8d1250c8a56c (cgit) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:15:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0e8cb748-e270-4d9d-9e6d-8d1250c8a56c is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f22e88-54cd-4802-bc55-577a23115959 (dotfiles) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f22e88-54cd-4802-bc55-577a23115959 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 39bcad9c-bdae-4e22-b89c-b6f057833c05 (homer) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 39bcad9c-bdae-4e22-b89c-b6f057833c05 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4212e8d9-a5f3-4e44-a65a-eec554ca52dc (phpmyadmin) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4212e8d9-a5f3-4e44-a65a-eec554ca52dc is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b5a31f83-3df5-4041-9b65-a060944e7586 (sosreport) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b5a31f83-3df5-4041-9b65-a060944e7586 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/deploy/500-kubernetes.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "++ semver 9.5.0 8.0.3",
                                "+ [[ 1 -ge 0 ]]",
                                "+ osism apply frr",
                                "\u001b[32m2026-03-26 02:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d590088-fe2b-4236-a786-40d95fe4bd4a (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d590088-fe2b-4236-a786-40d95fe4bd4a (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Thursday 26 March 2026  02:16:58 +0000 (0:00:00.547)       0:00:00.547 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Thursday 26 March 2026  02:16:59 +0000 (0:00:00.632)       0:00:01.180 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Thursday 26 March 2026  02:17:04 +0000 (0:00:05.690)       0:00:06.871 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Thursday 26 March 2026  02:17:17 +0000 (0:00:13.118)       0:00:19.989 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Thursday 26 March 2026  02:17:19 +0000 (0:00:01.409)       0:00:21.398 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Thursday 26 March 2026  02:17:21 +0000 (0:00:01.636)       0:00:23.035 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Thursday 26 March 2026  02:17:22 +0000 (0:00:01.493)       0:00:24.528 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Thursday 26 March 2026  02:17:22 +0000 (0:00:00.158)       0:00:24.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Thursday 26 March 2026  02:17:22 +0000 (0:00:00.193)       0:00:24.879 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Thursday 26 March 2026  02:17:23 +0000 (0:00:00.979)       0:00:25.859 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Thursday 26 March 2026  02:17:26 +0000 (0:00:02.627)       0:00:28.487 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.frr : Restart frr service] *********************",
                                "Thursday 26 March 2026  02:17:28 +0000 (0:00:02.013)       0:00:30.500 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:17:30 +0000 (0:00:01.549)       0:00:32.050 ******** ",
                                "=============================================================================== ",
                                "osism.services.frr : Install frr package ------------------------------- 13.12s",
                                "osism.services.frr : Pin frr package version ---------------------------- 5.69s",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 2.63s",
                                "osism.services.frr : Manage frr service --------------------------------- 2.01s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 1.64s",
                                "osism.services.frr : Restart frr service -------------------------------- 1.55s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.49s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.41s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.98s",
                                "osism.services.frr : Include distribution specific install tasks -------- 0.63s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 0.19s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 0.16s",
                                "+ osism apply kubernetes",
                                "\u001b[32m2026-03-26 02:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5f21ae79-fb4c-403b-8e3d-454a1387a6c5 (kubernetes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5f21ae79-fb4c-403b-8e3d-454a1387a6c5 (kubernetes) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Thursday 26 March 2026  02:17:38 +0000 (0:00:00.182)       0:00:00.182 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Thursday 26 March 2026  02:17:39 +0000 (0:00:00.753)       0:00:00.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Thursday 26 March 2026  02:17:40 +0000 (0:00:00.641)       0:00:01.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Thursday 26 March 2026  02:17:40 +0000 (0:00:00.818)       0:00:02.396 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Thursday 26 March 2026  02:17:42 +0000 (0:00:02.007)       0:00:04.403 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Thursday 26 March 2026  02:17:45 +0000 (0:00:02.155)       0:00:06.558 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Thursday 26 March 2026  02:17:46 +0000 (0:00:01.182)       0:00:07.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Thursday 26 March 2026  02:17:47 +0000 (0:00:00.898)       0:00:08.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Thursday 26 March 2026  02:17:47 +0000 (0:00:00.604)       0:00:09.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Thursday 26 March 2026  02:17:48 +0000 (0:00:00.709)       0:00:09.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Thursday 26 March 2026  02:17:49 +0000 (0:00:01.295)       0:00:11.249 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Thursday 26 March 2026  02:17:50 +0000 (0:00:00.855)       0:00:12.104 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Thursday 26 March 2026  02:17:55 +0000 (0:00:04.830)       0:00:16.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Thursday 26 March 2026  02:17:56 +0000 (0:00:01.074)       0:00:18.009 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Thursday 26 March 2026  02:17:57 +0000 (0:00:01.474)       0:00:19.483 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Thursday 26 March 2026  02:17:58 +0000 (0:00:00.776)       0:00:20.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Thursday 26 March 2026  02:17:59 +0000 (0:00:01.121)       0:00:21.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Thursday 26 March 2026  02:18:00 +0000 (0:00:00.675)       0:00:22.057 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Thursday 26 March 2026  02:18:01 +0000 (0:00:01.276)       0:00:23.334 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Thursday 26 March 2026  02:18:03 +0000 (0:00:01.765)       0:00:25.099 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Thursday 26 March 2026  02:18:05 +0000 (0:00:01.539)       0:00:26.639 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Thursday 26 March 2026  02:18:06 +0000 (0:00:01.156)       0:00:27.795 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Thursday 26 March 2026  02:18:07 +0000 (0:00:00.839)       0:00:28.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Thursday 26 March 2026  02:18:07 +0000 (0:00:00.419)       0:00:29.054 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Thursday 26 March 2026  02:18:08 +0000 (0:00:01.204)       0:00:30.259 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Thursday 26 March 2026  02:18:10 +0000 (0:00:01.634)       0:00:31.893 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Thursday 26 March 2026  02:18:11 +0000 (0:00:00.709)       0:00:32.603 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Thursday 26 March 2026  02:18:14 +0000 (0:00:02.900)       0:00:35.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Thursday 26 March 2026  02:18:14 +0000 (0:00:00.553)       0:00:36.056 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Thursday 26 March 2026  02:18:15 +0000 (0:00:00.796)       0:00:36.853 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Thursday 26 March 2026  02:18:16 +0000 (0:00:01.279)       0:00:38.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Thursday 26 March 2026  02:18:17 +0000 (0:00:00.802)       0:00:38.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Thursday 26 March 2026  02:18:17 +0000 (0:00:00.398)       0:00:39.334 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Thursday 26 March 2026  02:18:18 +0000 (0:00:01.123)       0:00:40.457 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Thursday 26 March 2026  02:18:22 +0000 (0:00:03.250)       0:00:43.707 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Thursday 26 March 2026  02:18:22 +0000 (0:00:00.530)       0:00:44.238 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Thursday 26 March 2026  02:19:16 +0000 (0:00:54.114)       0:01:38.353 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Thursday 26 March 2026  02:19:17 +0000 (0:00:00.342)       0:01:38.695 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Thursday 26 March 2026  02:19:18 +0000 (0:00:00.993)       0:01:39.689 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Thursday 26 March 2026  02:19:19 +0000 (0:00:01.187)       0:01:40.877 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Thursday 26 March 2026  02:19:46 +0000 (0:00:27.412)       0:02:08.290 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Thursday 26 March 2026  02:19:47 +0000 (0:00:00.649)       0:02:08.940 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Thursday 26 March 2026  02:19:48 +0000 (0:00:00.639)       0:02:09.580 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Thursday 26 March 2026  02:19:48 +0000 (0:00:00.667)       0:02:10.248 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Thursday 26 March 2026  02:19:49 +0000 (0:00:00.819)       0:02:11.067 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Thursday 26 March 2026  02:19:49 +0000 (0:00:00.332)       0:02:11.400 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Thursday 26 March 2026  02:19:50 +0000 (0:00:00.665)       0:02:12.066 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Thursday 26 March 2026  02:19:51 +0000 (0:00:00.672)       0:02:12.739 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Thursday 26 March 2026  02:19:52 +0000 (0:00:01.122)       0:02:13.861 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Thursday 26 March 2026  02:19:53 +0000 (0:00:00.881)       0:02:14.742 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Thursday 26 March 2026  02:19:53 +0000 (0:00:00.310)       0:02:15.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Thursday 26 March 2026  02:19:53 +0000 (0:00:00.344)       0:02:15.397 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Thursday 26 March 2026  02:19:54 +0000 (0:00:00.639)       0:02:16.037 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Thursday 26 March 2026  02:19:55 +0000 (0:00:00.907)       0:02:16.944 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Thursday 26 March 2026  02:19:58 +0000 (0:00:03.018)       0:02:19.963 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Thursday 26 March 2026  02:19:58 +0000 (0:00:00.350)       0:02:20.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Thursday 26 March 2026  02:19:59 +0000 (0:00:00.880)       0:02:21.195 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Thursday 26 March 2026  02:20:00 +0000 (0:00:00.323)       0:02:21.518 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Thursday 26 March 2026  02:20:00 +0000 (0:00:00.493)       0:02:22.012 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Thursday 26 March 2026  02:20:01 +0000 (0:00:00.563)       0:02:22.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Thursday 26 March 2026  02:20:01 +0000 (0:00:00.353)       0:02:22.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Thursday 26 March 2026  02:20:01 +0000 (0:00:00.325)       0:02:23.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Thursday 26 March 2026  02:20:02 +0000 (0:00:00.631)       0:02:23.886 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Thursday 26 March 2026  02:20:03 +0000 (0:00:01.410)       0:02:25.296 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Thursday 26 March 2026  02:20:05 +0000 (0:00:01.225)       0:02:26.522 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Thursday 26 March 2026  02:20:15 +0000 (0:00:10.243)       0:02:36.766 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Thursday 26 March 2026  02:20:16 +0000 (0:00:01.053)       0:02:37.820 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Thursday 26 March 2026  02:20:16 +0000 (0:00:00.471)       0:02:38.291 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Thursday 26 March 2026  02:20:17 +0000 (0:00:00.556)       0:02:38.848 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Thursday 26 March 2026  02:20:18 +0000 (0:00:00.897)       0:02:39.745 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Thursday 26 March 2026  02:20:18 +0000 (0:00:00.645)       0:02:40.390 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Thursday 26 March 2026  02:20:20 +0000 (0:00:01.629)       0:02:42.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Thursday 26 March 2026  02:20:21 +0000 (0:00:00.880)       0:02:42.901 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Thursday 26 March 2026  02:20:21 +0000 (0:00:00.494)       0:02:43.396 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Thursday 26 March 2026  02:20:22 +0000 (0:00:00.485)       0:02:43.881 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Thursday 26 March 2026  02:20:22 +0000 (0:00:00.393)       0:02:44.275 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Thursday 26 March 2026  02:20:23 +0000 (0:00:00.262)       0:02:44.537 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Thursday 26 March 2026  02:20:23 +0000 (0:00:00.867)       0:02:45.405 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Thursday 26 March 2026  02:20:25 +0000 (0:00:01.679)       0:02:47.085 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Thursday 26 March 2026  02:20:26 +0000 (0:00:00.838)       0:02:47.923 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Thursday 26 March 2026  02:20:26 +0000 (0:00:00.493)       0:02:48.416 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Thursday 26 March 2026  02:20:34 +0000 (0:00:07.985)       0:02:56.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Thursday 26 March 2026  02:20:47 +0000 (0:00:12.989)       0:03:09.392 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Thursday 26 March 2026  02:20:48 +0000 (0:00:00.817)       0:03:10.209 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Thursday 26 March 2026  02:20:49 +0000 (0:00:00.328)       0:03:10.538 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Thursday 26 March 2026  02:20:49 +0000 (0:00:00.326)       0:03:10.865 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Thursday 26 March 2026  02:20:50 +0000 (0:00:00.775)       0:03:11.640 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Thursday 26 March 2026  02:20:51 +0000 (0:00:00.915)       0:03:12.556 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Thursday 26 March 2026  02:20:52 +0000 (0:00:01.096)       0:03:13.653 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Thursday 26 March 2026  02:20:52 +0000 (0:00:00.134)       0:03:13.787 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Thursday 26 March 2026  02:20:53 +0000 (0:00:01.046)       0:03:14.834 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Thursday 26 March 2026  02:20:53 +0000 (0:00:00.142)       0:03:14.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Thursday 26 March 2026  02:20:53 +0000 (0:00:00.115)       0:03:15.092 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Thursday 26 March 2026  02:20:53 +0000 (0:00:00.131)       0:03:15.224 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Thursday 26 March 2026  02:20:53 +0000 (0:00:00.122)       0:03:15.346 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Thursday 26 March 2026  02:20:59 +0000 (0:00:05.567)       0:03:20.914 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Thursday 26 March 2026  02:21:41 +0000 (0:00:42.445)       0:04:03.359 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Thursday 26 March 2026  02:21:43 +0000 (0:00:01.296)       0:04:04.656 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Thursday 26 March 2026  02:21:44 +0000 (0:00:01.828)       0:04:06.484 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Thursday 26 March 2026  02:21:46 +0000 (0:00:01.154)       0:04:07.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Thursday 26 March 2026  02:21:46 +0000 (0:00:00.135)       0:04:07.774 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Thursday 26 March 2026  02:21:48 +0000 (0:00:01.862)       0:04:09.636 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Thursday 26 March 2026  02:21:48 +0000 (0:00:00.359)       0:04:09.996 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Thursday 26 March 2026  02:21:49 +0000 (0:00:00.850)       0:04:10.847 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Thursday 26 March 2026  02:21:49 +0000 (0:00:00.388)       0:04:11.235 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Thursday 26 March 2026  02:21:49 +0000 (0:00:00.237)       0:04:11.472 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Thursday 26 March 2026  02:21:55 +0000 (0:00:05.469)       0:04:16.942 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Thursday 26 March 2026  02:21:56 +0000 (0:00:00.807)       0:04:17.750 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Thursday 26 March 2026  02:22:04 +0000 (0:00:07.998)       0:04:25.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Thursday 26 March 2026  02:22:04 +0000 (0:00:00.578)       0:04:26.327 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:22:05 +0000 (0:00:00.706)       0:04:27.033 ******** ",
                                "=============================================================================== ",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 54.12s",
                                "k3s_server_post : Wait for Cilium resources ---------------------------- 42.45s",
                                "k3s_server : Enable and check K3s service ------------------------------ 27.41s",
                                "kubectl : Install required packages ------------------------------------ 12.99s",
                                "k3s_agent : Manage k3s service ----------------------------------------- 10.24s",
                                "Manage labels ----------------------------------------------------------- 8.00s",
                                "kubectl : Add repository Debian ----------------------------------------- 7.99s",
                                "k3s_server_post : Install Cilium ---------------------------------------- 5.57s",
                                "k9s : Install k9s packages ---------------------------------------------- 5.47s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 4.83s",
                                "k3s_server : Detect Kubernetes version for label compatibility ---------- 3.25s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 3.02s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.90s",
                                "k3s_prereq : Enable IPv6 forwarding ------------------------------------- 2.15s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.01s",
                                "k3s_server_post : Test for BGP config resources ------------------------- 1.86s",
                                "k3s_server_post : Copy BGP manifests to first master -------------------- 1.83s",
                                "k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers --- 1.77s",
                                "kubectl : Install apt-transport-https package --------------------------- 1.68s",
                                "k3s_server : Create custom resolv.conf for k3s -------------------------- 1.63s",
                                "+ osism apply copy-kubeconfig",
                                "\u001b[32m2026-03-26 02:22:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33d65d64-967a-44bc-a1e2-7e6dd84a9f9b (copy-kubeconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:22:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 33d65d64-967a-44bc-a1e2-7e6dd84a9f9b (copy-kubeconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy kubeconfig to the configuration repository] *************************",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Thursday 26 March 2026  02:22:22 +0000 (0:00:00.158)       0:00:00.158 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Thursday 26 March 2026  02:22:23 +0000 (0:00:00.792)       0:00:00.950 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig file] ****************************",
                                "Thursday 26 March 2026  02:22:24 +0000 (0:00:01.278)       0:00:02.229 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:22:25 +0000 (0:00:00.497)       0:00:02.727 ******** ",
                                "=============================================================================== ",
                                "Write kubeconfig file --------------------------------------------------- 1.28s",
                                "Get kubeconfig file ----------------------------------------------------- 0.79s",
                                "Change server address in the kubeconfig file ---------------------------- 0.50s",
                                "+ sh -c /opt/configuration/scripts/deploy/200-infrastructure.sh",
                                "\u001b[32m2026-03-26 02:22:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8310beb1-1a99-4254-99a6-2a442c4f14cd (openstackclient) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:22:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8310beb1-1a99-4254-99a6-2a442c4f14cd (openstackclient) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Thursday 26 March 2026  02:22:42 +0000 (0:00:00.237)       0:00:00.237 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Thursday 26 March 2026  02:22:43 +0000 (0:00:00.248)       0:00:00.485 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Thursday 26 March 2026  02:22:44 +0000 (0:00:01.298)       0:00:01.784 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Thursday 26 March 2026  02:22:45 +0000 (0:00:01.495)       0:00:03.279 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Thursday 26 March 2026  02:23:21 +0000 (0:00:35.575)       0:00:38.855 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Thursday 26 March 2026  02:23:22 +0000 (0:00:00.990)       0:00:39.845 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Thursday 26 March 2026  02:23:23 +0000 (0:00:00.734)       0:00:40.579 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Thursday 26 March 2026  02:23:24 +0000 (0:00:01.728)       0:00:42.308 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Thursday 26 March 2026  02:23:25 +0000 (0:00:00.772)       0:00:43.081 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Thursday 26 March 2026  02:23:26 +0000 (0:00:00.678)       0:00:43.760 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:23:26 +0000 (0:00:00.423)       0:00:44.183 ******** ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Manage openstackclient service -------- 35.58s",
                                "osism.services.openstackclient : Restart openstackclient service -------- 1.73s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 1.50s",
                                "osism.services.openstackclient : Create required directories ------------ 1.30s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 0.99s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 0.77s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 0.73s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.68s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 0.42s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.25s",
                                "\u001b[32m2026-03-26 02:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a0a4ffc3-3e59-4deb-90a6-de603effbb9d (common) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a0a4ffc3-3e59-4deb-90a6-de603effbb9d (common) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  02:23:34 +0000 (0:00:00.291)       0:00:00.291 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  02:23:35 +0000 (0:00:01.481)       0:00:01.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  02:23:38 +0000 (0:00:02.602)       0:00:04.375 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  02:23:39 +0000 (0:00:01.386)       0:00:05.761 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:23:43 +0000 (0:00:03.494)       0:00:09.256 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  02:23:44 +0000 (0:00:00.918)       0:00:10.174 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Thursday 26 March 2026  02:23:46 +0000 (0:00:02.013)       0:00:12.188 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Thursday 26 March 2026  02:23:46 +0000 (0:00:00.719)       0:00:12.907 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  02:23:47 +0000 (0:00:00.943)       0:00:13.850 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Thursday 26 March 2026  02:23:51 +0000 (0:00:03.512)       0:00:17.363 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Thursday 26 March 2026  02:23:52 +0000 (0:00:00.993)       0:00:18.357 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Thursday 26 March 2026  02:23:53 +0000 (0:00:01.351)       0:00:19.709 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Thursday 26 March 2026  02:23:54 +0000 (0:00:00.865)       0:00:20.574 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Thursday 26 March 2026  02:23:55 +0000 (0:00:00.949)       0:00:21.524 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Thursday 26 March 2026  02:23:58 +0000 (0:00:02.586)       0:00:24.110 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Thursday 26 March 2026  02:24:00 +0000 (0:00:02.216)       0:00:26.326 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Thursday 26 March 2026  02:24:02 +0000 (0:00:02.051)       0:00:28.377 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Thursday 26 March 2026  02:24:03 +0000 (0:00:01.523)       0:00:29.901 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Thursday 26 March 2026  02:24:05 +0000 (0:00:01.957)       0:00:31.859 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [common : Check common containers] ****************************************",
                                "Thursday 26 March 2026  02:24:07 +0000 (0:00:01.692)       0:00:33.552 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Creating log volume] ********************************************",
                                "Thursday 26 March 2026  02:24:10 +0000 (0:00:02.613)       0:00:36.166 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Link kolla_logs volume to /var/log/kolla] ***********************",
                                "Thursday 26 March 2026  02:24:11 +0000 (0:00:01.411)       0:00:37.577 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:12 +0000 (0:00:01.071)       0:00:38.648 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:12 +0000 (0:00:00.065)       0:00:38.713 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:12 +0000 (0:00:00.066)       0:00:38.780 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:12 +0000 (0:00:00.088)       0:00:38.869 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:13 +0000 (0:00:00.252)       0:00:39.121 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:13 +0000 (0:00:00.067)       0:00:39.188 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:24:13 +0000 (0:00:00.059)       0:00:39.248 ******** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Thursday 26 March 2026  02:24:13 +0000 (0:00:00.090)       0:00:39.338 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Thursday 26 March 2026  02:24:53 +0000 (0:00:40.562)       0:01:19.900 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Thursday 26 March 2026  02:25:25 +0000 (0:00:31.651)       0:01:51.552 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Thursday 26 March 2026  02:25:27 +0000 (0:00:01.878)       0:01:53.430 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:25:35 +0000 (0:00:08.580)       0:02:02.011 ******** ",
                                "=============================================================================== ",
                                "common : Restart fluentd container ------------------------------------- 40.56s",
                                "common : Restart kolla-toolbox container ------------------------------- 31.65s",
                                "common : Restart cron container ----------------------------------------- 8.58s",
                                "common : Copying over config.json files for services -------------------- 3.51s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 3.49s",
                                "common : Check common containers ---------------------------------------- 2.61s",
                                "common : Ensuring config directories exist ------------------------------ 2.60s",
                                "common : Copying over fluentd.conf -------------------------------------- 2.59s",
                                "common : Copying over cron logrotate config file ------------------------ 2.22s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.05s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 2.01s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 1.96s",
                                "common : Initializing toolbox container using normal user --------------- 1.88s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 1.69s",
                                "common : Ensuring config directories have correct owner and permission --- 1.52s",
                                "common : include_tasks -------------------------------------------------- 1.48s",
                                "common : Creating log volume -------------------------------------------- 1.41s",
                                "common : include_tasks -------------------------------------------------- 1.39s",
                                "common : Find custom fluentd filter config files ------------------------ 1.35s",
                                "common : Link kolla_logs volume to /var/log/kolla ----------------------- 1.07s",
                                "\u001b[32m2026-03-26 02:25:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69ee4675-db84-4d53-8664-3bcc753c27a0 (loadbalancer) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:25:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 69ee4675-db84-4d53-8664-3bcc753c27a0 (loadbalancer) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:25:43 +0000 (0:00:00.271)       0:00:00.271 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:25:43 +0000 (0:00:00.310)       0:00:00.582 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Thursday 26 March 2026  02:25:44 +0000 (0:00:00.455)       0:00:01.037 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Thursday 26 March 2026  02:25:44 +0000 (0:00:00.598)       0:00:01.635 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Thursday 26 March 2026  02:25:45 +0000 (0:00:00.619)       0:00:02.255 ******** ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Thursday 26 March 2026  02:25:45 +0000 (0:00:00.735)       0:00:02.991 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Thursday 26 March 2026  02:25:46 +0000 (0:00:00.601)       0:00:03.593 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Thursday 26 March 2026  02:25:48 +0000 (0:00:02.236)       0:00:05.829 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Thursday 26 March 2026  02:25:49 +0000 (0:00:00.742)       0:00:06.572 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Thursday 26 March 2026  02:25:50 +0000 (0:00:01.251)       0:00:07.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Thursday 26 March 2026  02:25:51 +0000 (0:00:00.553)       0:00:08.376 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Thursday 26 March 2026  02:25:53 +0000 (0:00:01.836)       0:00:10.213 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Thursday 26 March 2026  02:25:54 +0000 (0:00:00.910)       0:00:11.124 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Thursday 26 March 2026  02:25:55 +0000 (0:00:01.544)       0:00:12.669 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Thursday 26 March 2026  02:25:56 +0000 (0:00:00.931)       0:00:13.600 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Thursday 26 March 2026  02:25:57 +0000 (0:00:01.366)       0:00:14.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Thursday 26 March 2026  02:25:58 +0000 (0:00:00.605)       0:00:15.572 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.6.20251130', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a', '__omit_place_holder__5801856146be98cdce07043bf9184b3fff60071a'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Thursday 26 March 2026  02:26:01 +0000 (0:00:02.830)       0:00:18.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Thursday 26 March 2026  02:26:04 +0000 (0:00:03.217)       0:00:21.620 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Thursday 26 March 2026  02:26:06 +0000 (0:00:01.956)       0:00:23.576 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Thursday 26 March 2026  02:26:09 +0000 (0:00:03.037)       0:00:26.614 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Thursday 26 March 2026  02:26:10 +0000 (0:00:00.609)       0:00:27.223 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Thursday 26 March 2026  02:26:12 +0000 (0:00:02.106)       0:00:29.329 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Thursday 26 March 2026  02:26:14 +0000 (0:00:02.125)       0:00:31.455 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Thursday 26 March 2026  02:26:15 +0000 (0:00:01.462)       0:00:32.918 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Thursday 26 March 2026  02:26:17 +0000 (0:00:01.414)       0:00:34.332 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***",
                                "Thursday 26 March 2026  02:26:17 +0000 (0:00:00.556)       0:00:34.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:26:21 +0000 (0:00:03.351)       0:00:38.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  02:26:21 +0000 (0:00:00.620)       0:00:38.861 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  02:26:22 +0000 (0:00:00.800)       0:00:39.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:26:23 +0000 (0:00:01.018)       0:00:40.680 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  02:26:24 +0000 (0:00:00.654)       0:00:41.334 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******",
                                "Thursday 26 March 2026  02:26:25 +0000 (0:00:00.822)       0:00:42.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:26:26 +0000 (0:00:01.113)       0:00:43.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****",
                                "Thursday 26 March 2026  02:26:26 +0000 (0:00:00.614)       0:00:43.885 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Thursday 26 March 2026  02:26:27 +0000 (0:00:00.804)       0:00:44.689 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Thursday 26 March 2026  02:26:29 +0000 (0:00:01.699)       0:00:46.389 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Thursday 26 March 2026  02:26:31 +0000 (0:00:01.762)       0:00:48.152 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check loadbalancer containers] ****************************",
                                "Thursday 26 March 2026  02:26:31 +0000 (0:00:00.812)       0:00:48.964 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.8.15.20251130', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:3.0.3.20251130', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.8.20251130', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Thursday 26 March 2026  02:26:34 +0000 (0:00:02.431)       0:00:51.395 ******** ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Thursday 26 March 2026  02:26:35 +0000 (0:00:00.818)       0:00:52.213 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Thursday 26 March 2026  02:26:38 +0000 (0:00:03.286)       0:00:55.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Thursday 26 March 2026  02:26:39 +0000 (0:00:00.660)       0:00:56.161 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Thursday 26 March 2026  02:26:40 +0000 (0:00:01.127)       0:00:57.288 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Thursday 26 March 2026  02:26:41 +0000 (0:00:01.295)       0:00:58.584 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Thursday 26 March 2026  02:26:43 +0000 (0:00:02.068)       0:01:00.653 ******** ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Thursday 26 March 2026  02:26:44 +0000 (0:00:00.678)       0:01:01.331 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Thursday 26 March 2026  02:26:47 +0000 (0:00:03.402)       0:01:04.734 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Thursday 26 March 2026  02:26:48 +0000 (0:00:00.668)       0:01:05.403 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Thursday 26 March 2026  02:26:49 +0000 (0:00:00.828)       0:01:06.232 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  02:26:50 +0000 (0:00:01.596)       0:01:07.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Thursday 26 March 2026  02:26:52 +0000 (0:00:02.053)       0:01:09.882 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Thursday 26 March 2026  02:26:53 +0000 (0:00:00.333)       0:01:10.215 ******** ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Thursday 26 March 2026  02:26:53 +0000 (0:00:00.738)       0:01:10.954 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Thursday 26 March 2026  02:26:56 +0000 (0:00:02.822)       0:01:13.776 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Thursday 26 March 2026  02:26:58 +0000 (0:00:01.575)       0:01:15.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Thursday 26 March 2026  02:26:59 +0000 (0:00:01.637)       0:01:16.990 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  02:27:00 +0000 (0:00:00.440)       0:01:17.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Thursday 26 March 2026  02:27:01 +0000 (0:00:01.296)       0:01:18.727 ******** ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Thursday 26 March 2026  02:27:02 +0000 (0:00:01.012)       0:01:19.740 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Thursday 26 March 2026  02:27:06 +0000 (0:00:03.487)       0:01:23.227 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Thursday 26 March 2026  02:27:06 +0000 (0:00:00.759)       0:01:23.986 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Thursday 26 March 2026  02:27:08 +0000 (0:00:01.222)       0:01:25.209 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Thursday 26 March 2026  02:27:09 +0000 (0:00:01.266)       0:01:26.476 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Thursday 26 March 2026  02:27:11 +0000 (0:00:02.038)       0:01:28.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Thursday 26 March 2026  02:27:11 +0000 (0:00:00.313)       0:01:28.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Thursday 26 March 2026  02:27:12 +0000 (0:00:00.338)       0:01:29.167 ******** ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Thursday 26 March 2026  02:27:13 +0000 (0:00:00.976)       0:01:30.143 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Thursday 26 March 2026  02:27:16 +0000 (0:00:03.718)       0:01:33.861 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Thursday 26 March 2026  02:27:17 +0000 (0:00:01.078)       0:01:34.940 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:27:19 +0000 (0:00:01.314)       0:01:36.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:27:20 +0000 (0:00:01.332)       0:01:37.587 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Thursday 26 March 2026  02:27:22 +0000 (0:00:02.013)       0:01:39.600 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Thursday 26 March 2026  02:27:22 +0000 (0:00:00.354)       0:01:39.955 ******** ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Thursday 26 March 2026  02:27:24 +0000 (0:00:01.159)       0:01:41.114 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Thursday 26 March 2026  02:27:28 +0000 (0:00:04.520)       0:01:45.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20251130', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Thursday 26 March 2026  02:27:31 +0000 (0:00:02.970)       0:01:48.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Thursday 26 March 2026  02:27:35 +0000 (0:00:04.351)       0:01:52.957 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Thursday 26 March 2026  02:27:37 +0000 (0:00:01.349)       0:01:54.307 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Thursday 26 March 2026  02:27:39 +0000 (0:00:02.134)       0:01:56.441 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Thursday 26 March 2026  02:27:39 +0000 (0:00:00.359)       0:01:56.800 ******** ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Thursday 26 March 2026  02:27:40 +0000 (0:00:01.123)       0:01:57.924 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Thursday 26 March 2026  02:27:43 +0000 (0:00:02.977)       0:02:00.901 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Thursday 26 March 2026  02:27:44 +0000 (0:00:00.400)       0:02:01.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:27:45 +0000 (0:00:00.930)       0:02:02.232 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:27:46 +0000 (0:00:01.294)       0:02:03.527 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Thursday 26 March 2026  02:27:48 +0000 (0:00:02.027)       0:02:05.555 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Thursday 26 March 2026  02:27:48 +0000 (0:00:00.342)       0:02:05.897 ******** ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Thursday 26 March 2026  02:27:50 +0000 (0:00:01.208)       0:02:07.106 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Thursday 26 March 2026  02:27:53 +0000 (0:00:03.313)       0:02:10.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Thursday 26 March 2026  02:27:54 +0000 (0:00:00.742)       0:02:11.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:27:55 +0000 (0:00:01.167)       0:02:12.330 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:27:56 +0000 (0:00:01.631)       0:02:13.962 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Thursday 26 March 2026  02:27:59 +0000 (0:00:02.205)       0:02:16.167 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Thursday 26 March 2026  02:27:59 +0000 (0:00:00.310)       0:02:16.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Thursday 26 March 2026  02:27:59 +0000 (0:00:00.346)       0:02:16.824 ******** ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Thursday 26 March 2026  02:28:01 +0000 (0:00:01.252)       0:02:18.077 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Thursday 26 March 2026  02:28:04 +0000 (0:00:03.224)       0:02:21.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Thursday 26 March 2026  02:28:04 +0000 (0:00:00.702)       0:02:22.004 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Thursday 26 March 2026  02:28:06 +0000 (0:00:01.183)       0:02:23.187 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  02:28:07 +0000 (0:00:01.380)       0:02:24.568 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Thursday 26 March 2026  02:28:09 +0000 (0:00:02.118)       0:02:26.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Thursday 26 March 2026  02:28:09 +0000 (0:00:00.315)       0:02:27.002 ******** ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Thursday 26 March 2026  02:28:11 +0000 (0:00:01.413)       0:02:28.415 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Thursday 26 March 2026  02:28:14 +0000 (0:00:03.550)       0:02:31.966 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Thursday 26 March 2026  02:28:15 +0000 (0:00:00.753)       0:02:32.720 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Thursday 26 March 2026  02:28:16 +0000 (0:00:00.941)       0:02:33.661 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Thursday 26 March 2026  02:28:18 +0000 (0:00:01.687)       0:02:35.349 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Thursday 26 March 2026  02:28:20 +0000 (0:00:02.132)       0:02:37.481 ******** ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Thursday 26 March 2026  02:28:21 +0000 (0:00:01.120)       0:02:38.601 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Thursday 26 March 2026  02:28:25 +0000 (0:00:03.958)       0:02:42.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Thursday 26 March 2026  02:28:26 +0000 (0:00:01.091)       0:02:43.651 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Thursday 26 March 2026  02:28:27 +0000 (0:00:00.947)       0:02:44.599 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Thursday 26 March 2026  02:28:28 +0000 (0:00:01.299)       0:02:45.899 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Thursday 26 March 2026  02:28:31 +0000 (0:00:02.141)       0:02:48.041 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Thursday 26 March 2026  02:28:32 +0000 (0:00:01.387)       0:02:49.428 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Thursday 26 March 2026  02:28:35 +0000 (0:00:03.120)       0:02:52.548 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Thursday 26 March 2026  02:28:38 +0000 (0:00:02.467)       0:02:55.015 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Thursday 26 March 2026  02:28:40 +0000 (0:00:02.500)       0:02:57.516 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:28:43 +0000 (0:00:03.062)       0:03:00.579 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:28:45 +0000 (0:00:02.181)       0:03:02.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Thursday 26 March 2026  02:28:47 +0000 (0:00:01.460)       0:03:04.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Thursday 26 March 2026  02:28:47 +0000 (0:00:00.346)       0:03:04.568 ******** ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Thursday 26 March 2026  02:28:49 +0000 (0:00:01.452)       0:03:06.021 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Thursday 26 March 2026  02:28:50 +0000 (0:00:01.598)       0:03:07.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.24.20251130', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Thursday 26 March 2026  02:28:51 +0000 (0:00:00.405)       0:03:08.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:28:51 +0000 (0:00:00.952)       0:03:08.977 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:28:52 +0000 (0:00:00.443)       0:03:09.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Thursday 26 March 2026  02:28:53 +0000 (0:00:01.457)       0:03:10.878 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Thursday 26 March 2026  02:28:54 +0000 (0:00:00.351)       0:03:11.230 ******** ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Thursday 26 March 2026  02:28:55 +0000 (0:00:01.591)       0:03:12.822 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:00 +0000 (0:00:04.359)       0:03:17.182 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.2.20251130', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.2.20251130', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.2.20251130', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.2.20251130', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.2.20251130', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Thursday 26 March 2026  02:29:01 +0000 (0:00:01.524)       0:03:18.706 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:29:03 +0000 (0:00:02.029)       0:03:20.736 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:29:05 +0000 (0:00:01.377)       0:03:22.113 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Thursday 26 March 2026  02:29:07 +0000 (0:00:02.167)       0:03:24.281 ******** ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Thursday 26 March 2026  02:29:08 +0000 (0:00:01.251)       0:03:25.533 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:12 +0000 (0:00:03.611)       0:03:29.144 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Thursday 26 March 2026  02:29:12 +0000 (0:00:00.545)       0:03:29.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:29:13 +0000 (0:00:00.799)       0:03:30.490 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:29:15 +0000 (0:00:01.966)       0:03:32.457 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Thursday 26 March 2026  02:29:17 +0000 (0:00:01.930)       0:03:34.387 ******** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Thursday 26 March 2026  02:29:18 +0000 (0:00:01.583)       0:03:35.971 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:23 +0000 (0:00:04.202)       0:03:40.173 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Thursday 26 March 2026  02:29:24 +0000 (0:00:01.431)       0:03:41.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Thursday 26 March 2026  02:29:25 +0000 (0:00:01.078)       0:03:42.683 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Thursday 26 March 2026  02:29:27 +0000 (0:00:01.507)       0:03:44.190 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Thursday 26 March 2026  02:29:29 +0000 (0:00:02.166)       0:03:46.356 ******** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Thursday 26 March 2026  02:29:31 +0000 (0:00:01.659)       0:03:48.016 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Thursday 26 March 2026  02:29:31 +0000 (0:00:00.881)       0:03:48.897 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:35 +0000 (0:00:04.092)       0:03:52.990 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Thursday 26 March 2026  02:29:37 +0000 (0:00:01.469)       0:03:54.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:29:38 +0000 (0:00:01.527)       0:03:55.987 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:29:41 +0000 (0:00:02.483)       0:03:58.470 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Thursday 26 March 2026  02:29:44 +0000 (0:00:02.870)       0:04:01.341 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Thursday 26 March 2026  02:29:45 +0000 (0:00:01.117)       0:04:02.458 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:46 +0000 (0:00:01.039)       0:04:03.498 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Thursday 26 March 2026  02:29:47 +0000 (0:00:01.291)       0:04:04.789 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:29:49 +0000 (0:00:01.606)       0:04:06.396 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:29:52 +0000 (0:00:02.774)       0:04:09.171 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Thursday 26 March 2026  02:29:54 +0000 (0:00:02.714)       0:04:11.885 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Thursday 26 March 2026  02:29:56 +0000 (0:00:01.265)       0:04:13.151 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Thursday 26 March 2026  02:29:57 +0000 (0:00:01.376)       0:04:14.527 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Thursday 26 March 2026  02:29:58 +0000 (0:00:01.422)       0:04:15.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  02:30:00 +0000 (0:00:01.896)       0:04:17.847 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  02:30:04 +0000 (0:00:03.266)       0:04:21.114 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Thursday 26 March 2026  02:30:07 +0000 (0:00:03.279)       0:04:24.394 ******** ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Thursday 26 March 2026  02:30:08 +0000 (0:00:01.483)       0:04:25.878 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Thursday 26 March 2026  02:30:12 +0000 (0:00:03.331)       0:04:29.209 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Thursday 26 March 2026  02:30:12 +0000 (0:00:00.741)       0:04:29.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:30:13 +0000 (0:00:00.964)       0:04:30.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:30:15 +0000 (0:00:01.802)       0:04:32.717 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Thursday 26 March 2026  02:30:17 +0000 (0:00:02.208)       0:04:34.926 ******** ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Thursday 26 March 2026  02:30:19 +0000 (0:00:01.424)       0:04:36.351 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Thursday 26 March 2026  02:30:24 +0000 (0:00:05.349)       0:04:41.700 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Thursday 26 March 2026  02:30:25 +0000 (0:00:01.069)       0:04:42.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Thursday 26 March 2026  02:30:26 +0000 (0:00:00.994)       0:04:43.764 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Thursday 26 March 2026  02:30:27 +0000 (0:00:00.451)       0:04:44.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Thursday 26 March 2026  02:30:28 +0000 (0:00:01.471)       0:04:45.688 ******** ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Thursday 26 March 2026  02:30:30 +0000 (0:00:01.846)       0:04:47.534 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Thursday 26 March 2026  02:30:34 +0000 (0:00:04.189)       0:04:51.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:1.7.0.20251130', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Thursday 26 March 2026  02:30:35 +0000 (0:00:00.854)       0:04:52.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Thursday 26 March 2026  02:30:37 +0000 (0:00:01.495)       0:04:54.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Thursday 26 March 2026  02:30:37 +0000 (0:00:00.480)       0:04:54.553 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Thursday 26 March 2026  02:30:38 +0000 (0:00:01.393)       0:04:55.946 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Thursday 26 March 2026  02:30:40 +0000 (0:00:01.863)       0:04:57.810 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Thursday 26 March 2026  02:30:43 +0000 (0:00:02.204)       0:05:00.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Thursday 26 March 2026  02:30:43 +0000 (0:00:00.433)       0:05:00.448 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Thursday 26 March 2026  02:30:44 +0000 (0:00:01.000)       0:05:01.448 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  02:30:44 +0000 (0:00:00.502)       0:05:01.951 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Thursday 26 March 2026  02:30:46 +0000 (0:00:01.439)       0:05:03.390 ******** ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Thursday 26 March 2026  02:30:47 +0000 (0:00:01.611)       0:05:05.002 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Thursday 26 March 2026  02:30:54 +0000 (0:00:06.936)       0:05:11.938 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Thursday 26 March 2026  02:30:55 +0000 (0:00:00.770)       0:05:12.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Thursday 26 March 2026  02:30:56 +0000 (0:00:00.933)       0:05:13.642 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Thursday 26 March 2026  02:30:58 +0000 (0:00:01.400)       0:05:15.043 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : swift] ****************************************************",
                                "Thursday 26 March 2026  02:31:00 +0000 (0:00:02.290)       0:05:17.333 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Thursday 26 March 2026  02:31:00 +0000 (0:00:00.665)       0:05:17.999 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Thursday 26 March 2026  02:31:01 +0000 (0:00:00.356)       0:05:18.356 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Thursday 26 March 2026  02:31:01 +0000 (0:00:00.438)       0:05:18.794 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Thursday 26 March 2026  02:31:02 +0000 (0:00:00.337)       0:05:19.132 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Thursday 26 March 2026  02:31:02 +0000 (0:00:00.680)       0:05:19.812 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Thursday 26 March 2026  02:31:03 +0000 (0:00:00.602)       0:05:20.414 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Thursday 26 March 2026  02:31:04 +0000 (0:00:00.716)       0:05:21.130 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Thursday 26 March 2026  02:31:04 +0000 (0:00:00.729)       0:05:21.860 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Thursday 26 March 2026  02:31:05 +0000 (0:00:00.892)       0:05:22.753 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Thursday 26 March 2026  02:31:06 +0000 (0:00:00.969)       0:05:23.722 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Thursday 26 March 2026  02:31:07 +0000 (0:00:00.842)       0:05:24.565 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Thursday 26 March 2026  02:31:12 +0000 (0:00:05.397)       0:05:29.962 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Thursday 26 March 2026  02:31:16 +0000 (0:00:03.222)       0:05:33.185 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Thursday 26 March 2026  02:31:27 +0000 (0:00:10.974)       0:05:44.159 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Thursday 26 March 2026  02:31:31 +0000 (0:00:04.647)       0:05:48.807 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Thursday 26 March 2026  02:31:36 +0000 (0:00:04.277)       0:05:53.085 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Thursday 26 March 2026  02:31:36 +0000 (0:00:00.761)       0:05:53.847 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Thursday 26 March 2026  02:31:37 +0000 (0:00:00.403)       0:05:54.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Thursday 26 March 2026  02:31:37 +0000 (0:00:00.388)       0:05:54.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Thursday 26 March 2026  02:31:38 +0000 (0:00:00.398)       0:05:55.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Thursday 26 March 2026  02:31:38 +0000 (0:00:00.737)       0:05:55.775 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Thursday 26 March 2026  02:31:39 +0000 (0:00:00.373)       0:05:56.148 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Thursday 26 March 2026  02:31:43 +0000 (0:00:04.783)       0:06:00.932 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:31:44 +0000 (0:00:00.819)       0:06:01.751 ******** ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup proxysql container ------------------------- 10.97s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 6.94s",
                                "loadbalancer : Start backup haproxy container --------------------------- 5.40s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 5.35s",
                                "loadbalancer : Wait for haproxy to listen on VIP ------------------------ 4.78s",
                                "loadbalancer : Wait for backup proxysql to start ------------------------ 4.65s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 4.52s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 4.36s",
                                "haproxy-config : Configuring firewall for glance ------------------------ 4.35s",
                                "loadbalancer : Start backup keepalived container ------------------------ 4.28s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 4.20s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 4.19s",
                                "haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.09s",
                                "haproxy-config : Copying over manila haproxy config --------------------- 3.96s",
                                "haproxy-config : Copying over designate haproxy config ------------------ 3.72s",
                                "haproxy-config : Copying over placement haproxy config ------------------ 3.61s",
                                "haproxy-config : Copying over magnum haproxy config --------------------- 3.55s",
                                "haproxy-config : Copying over cinder haproxy config --------------------- 3.49s",
                                "haproxy-config : Copying over barbican haproxy config ------------------- 3.40s",
                                "service-cert-copy : loadbalancer | Copying over extra CA certificates --- 3.35s",
                                "\u001b[32m2026-03-26 02:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d9981432-5afc-4da2-8289-2d0601443640 (opensearch) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d9981432-5afc-4da2-8289-2d0601443640 (opensearch) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:31:52 +0000 (0:00:00.269)       0:00:00.269 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:31:52 +0000 (0:00:00.298)       0:00:00.568 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  02:31:53 +0000 (0:00:00.447)       0:00:01.015 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Thursday 26 March 2026  02:31:53 +0000 (0:00:00.538)       0:00:01.554 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Thursday 26 March 2026  02:31:54 +0000 (0:00:00.667)       0:00:02.221 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  02:31:56 +0000 (0:00:01.732)       0:00:03.953 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Thursday 26 March 2026  02:31:56 +0000 (0:00:00.571)       0:00:04.525 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:31:59 +0000 (0:00:02.402)       0:00:06.927 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  02:32:00 +0000 (0:00:00.819)       0:00:07.747 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Thursday 26 March 2026  02:32:01 +0000 (0:00:01.091)       0:00:08.838 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Thursday 26 March 2026  02:32:03 +0000 (0:00:02.510)       0:00:11.349 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Thursday 26 March 2026  02:32:06 +0000 (0:00:02.379)       0:00:13.729 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Check opensearch containers] ********************************",
                                "Thursday 26 March 2026  02:32:07 +0000 (0:00:01.915)       0:00:15.644 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.4.20251130', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  02:32:09 +0000 (0:00:01.657)       0:00:17.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  02:32:09 +0000 (0:00:00.299)       0:00:17.601 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  02:32:10 +0000 (0:00:00.080)       0:00:17.682 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  02:32:10 +0000 (0:00:00.067)       0:00:17.750 ******** ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Thursday 26 March 2026  02:32:10 +0000 (0:00:00.090)       0:00:17.840 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Thursday 26 March 2026  02:32:10 +0000 (0:00:00.226)       0:00:18.066 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Thursday 26 March 2026  02:32:11 +0000 (0:00:00.683)       0:00:18.750 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Thursday 26 March 2026  02:33:22 +0000 (0:01:10.951)       0:01:29.701 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  02:34:46 +0000 (0:01:24.336)       0:02:54.038 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Thursday 26 March 2026  02:34:46 +0000 (0:00:00.522)       0:02:54.561 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Thursday 26 March 2026  02:34:49 +0000 (0:00:02.979)       0:02:57.540 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Thursday 26 March 2026  02:34:52 +0000 (0:00:02.245)       0:02:59.786 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Thursday 26 March 2026  02:34:54 +0000 (0:00:02.731)       0:03:02.518 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:34:57 +0000 (0:00:02.568)       0:03:05.086 ******** ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 84.34s",
                                "opensearch : Restart opensearch container ------------------------------ 70.95s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 2.98s",
                                "opensearch : Create new log retention policy ---------------------------- 2.73s",
                                "opensearch : Apply retention policy to existing indices ----------------- 2.57s",
                                "opensearch : Copying over config.json files for services ---------------- 2.51s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 2.40s",
                                "opensearch : Copying over opensearch service config file ---------------- 2.38s",
                                "opensearch : Check if a log retention policy exists --------------------- 2.25s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 1.92s",
                                "opensearch : Ensuring config directories exist -------------------------- 1.73s",
                                "opensearch : Check opensearch containers -------------------------------- 1.66s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 1.09s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 0.82s",
                                "opensearch : Perform a flush -------------------------------------------- 0.68s",
                                "opensearch : Setting sysctl values -------------------------------------- 0.67s",
                                "opensearch : include_tasks ---------------------------------------------- 0.57s",
                                "opensearch : include_tasks ---------------------------------------------- 0.54s",
                                "opensearch : include_tasks ---------------------------------------------- 0.52s",
                                "Group hosts based on enabled services ----------------------------------- 0.45s",
                                "\u001b[32m2026-03-26 02:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e3d112e1-7037-4de8-b783-8f3076c4424e (memcached) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e3d112e1-7037-4de8-b783-8f3076c4424e (memcached) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:35:04 +0000 (0:00:00.281)       0:00:00.281 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:35:04 +0000 (0:00:00.362)       0:00:00.644 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  02:35:05 +0000 (0:00:00.469)       0:00:01.113 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  02:35:05 +0000 (0:00:00.498)       0:00:01.611 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Thursday 26 March 2026  02:35:06 +0000 (0:00:00.681)       0:00:02.293 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Check memcached container] ***********************************",
                                "Thursday 26 March 2026  02:35:08 +0000 (0:00:01.774)       0:00:04.067 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Thursday 26 March 2026  02:35:09 +0000 (0:00:01.577)       0:00:05.645 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:35:12 +0000 (0:00:02.122)       0:00:07.767 ******** ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container --------------------------------- 2.12s",
                                "memcached : Copying over config.json files for services ----------------- 1.77s",
                                "memcached : Check memcached container ----------------------------------- 1.58s",
                                "memcached : Ensuring config directories exist --------------------------- 0.68s",
                                "memcached : include_tasks ----------------------------------------------- 0.50s",
                                "Group hosts based on enabled services ----------------------------------- 0.47s",
                                "Group hosts based on Kolla action --------------------------------------- 0.36s",
                                "\u001b[32m2026-03-26 02:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a877614-1e0e-4232-a1a6-b271c8628b0b (redis) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3a877614-1e0e-4232-a1a6-b271c8628b0b (redis) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:35:19 +0000 (0:00:00.275)       0:00:00.275 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:35:19 +0000 (0:00:00.333)       0:00:00.609 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Thursday 26 March 2026  02:35:20 +0000 (0:00:00.433)       0:00:01.042 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Thursday 26 March 2026  02:35:20 +0000 (0:00:00.491)       0:00:01.534 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Thursday 26 March 2026  02:35:21 +0000 (0:00:01.076)       0:00:02.610 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Thursday 26 March 2026  02:35:24 +0000 (0:00:02.523)       0:00:05.134 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Check redis containers] ******************************************",
                                "Thursday 26 March 2026  02:35:26 +0000 (0:00:02.558)       0:00:07.692 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20251130', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  02:35:28 +0000 (0:00:01.467)       0:00:09.159 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  02:35:28 +0000 (0:00:00.068)       0:00:09.227 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  02:35:28 +0000 (0:00:00.069)       0:00:09.297 ******** ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Thursday 26 March 2026  02:35:28 +0000 (0:00:00.067)       0:00:09.364 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Thursday 26 March 2026  02:35:36 +0000 (0:00:07.569)       0:00:16.934 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:35:44 +0000 (0:00:08.438)       0:00:25.373 ******** ",
                                "=============================================================================== ",
                                "redis : Restart redis-sentinel container -------------------------------- 8.44s",
                                "redis : Restart redis container ----------------------------------------- 7.57s",
                                "redis : Copying over redis config files --------------------------------- 2.56s",
                                "redis : Copying over default config.json files -------------------------- 2.52s",
                                "redis : Check redis containers ------------------------------------------ 1.47s",
                                "redis : Ensuring config directories exist ------------------------------- 1.08s",
                                "redis : include_tasks --------------------------------------------------- 0.49s",
                                "Group hosts based on enabled services ----------------------------------- 0.43s",
                                "Group hosts based on Kolla action --------------------------------------- 0.33s",
                                "redis : Flush handlers -------------------------------------------------- 0.21s",
                                "\u001b[32m2026-03-26 02:35:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33359332-80a6-46d1-87b5-7308d2e3af07 (mariadb) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:35:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 33359332-80a6-46d1-87b5-7308d2e3af07 (mariadb) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:35:51 +0000 (0:00:00.167)       0:00:00.167 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:35:52 +0000 (0:00:00.346)       0:00:00.514 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Thursday 26 March 2026  02:35:52 +0000 (0:00:00.577)       0:00:01.091 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  02:35:53 +0000 (0:00:00.376)       0:00:01.468 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  02:35:53 +0000 (0:00:00.568)       0:00:02.036 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Thursday 26 March 2026  02:35:56 +0000 (0:00:02.874)       0:00:04.911 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Thursday 26 March 2026  02:35:57 +0000 (0:00:00.647)       0:00:05.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  02:35:58 +0000 (0:00:01.395)       0:00:06.954 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Thursday 26 March 2026  02:36:01 +0000 (0:00:03.262)       0:00:10.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Thursday 26 March 2026  02:36:02 +0000 (0:00:01.090)       0:00:11.307 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  02:36:06 +0000 (0:00:03.848)       0:00:15.156 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  02:36:07 +0000 (0:00:00.552)       0:00:15.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  02:36:09 +0000 (0:00:02.272)       0:00:17.980 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  02:36:12 +0000 (0:00:02.584)       0:00:20.565 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check mariadb containers] **************************************",
                                "Thursday 26 March 2026  02:36:14 +0000 (0:00:02.268)       0:00:22.833 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Thursday 26 March 2026  02:36:17 +0000 (0:00:02.683)       0:00:25.517 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Thursday 26 March 2026  02:36:17 +0000 (0:00:00.831)       0:00:26.348 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Thursday 26 March 2026  02:36:18 +0000 (0:00:00.560)       0:00:26.908 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Thursday 26 March 2026  02:36:18 +0000 (0:00:00.356)       0:00:27.265 ******** ",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Thursday 26 March 2026  02:36:29 +0000 (0:00:10.949)       0:00:38.215 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Thursday 26 March 2026  02:36:30 +0000 (0:00:00.426)       0:00:38.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Thursday 26 March 2026  02:36:30 +0000 (0:00:00.685)       0:00:39.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Thursday 26 March 2026  02:36:31 +0000 (0:00:00.441)       0:00:39.767 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Thursday 26 March 2026  02:36:31 +0000 (0:00:00.424)       0:00:40.192 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Thursday 26 March 2026  02:36:32 +0000 (0:00:00.431)       0:00:40.623 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  02:36:33 +0000 (0:00:00.915)       0:00:41.539 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [mariadb : Running MariaDB bootstrap container] ***************************",
                                "Thursday 26 March 2026  02:36:33 +0000 (0:00:00.400)       0:00:41.939 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap host name into facts] **************************",
                                "Thursday 26 March 2026  02:36:43 +0000 (0:00:10.184)       0:00:52.124 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  02:36:43 +0000 (0:00:00.133)       0:00:52.257 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Starting first MariaDB container] *******************",
                                "Thursday 26 March 2026  02:36:44 +0000 (0:00:01.086)       0:00:53.344 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******",
                                "Thursday 26 March 2026  02:36:52 +0000 (0:00:07.778)       0:01:01.123 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******",
                                "Thursday 26 March 2026  02:36:54 +0000 (0:00:01.580)       0:01:02.703 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***",
                                "Thursday 26 March 2026  02:36:56 +0000 (0:00:02.667)       0:01:05.370 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Thursday 26 March 2026  02:36:57 +0000 (0:00:00.144)       0:01:05.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Thursday 26 March 2026  02:36:57 +0000 (0:00:00.327)       0:01:05.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  02:36:57 +0000 (0:00:00.573)       0:01:06.416 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  02:37:16 +0000 (0:00:18.300)       0:01:24.717 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  02:37:32 +0000 (0:00:16.579)       0:01:41.296 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  02:37:35 +0000 (0:00:02.459)       0:01:43.756 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  02:37:53 +0000 (0:00:18.549)       0:02:02.306 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  02:38:10 +0000 (0:00:16.528)       0:02:18.834 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  02:38:12 +0000 (0:00:02.554)       0:02:21.389 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  02:38:25 +0000 (0:00:12.060)       0:02:33.449 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  02:38:29 +0000 (0:00:04.563)       0:02:38.012 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Thursday 26 March 2026  02:38:32 +0000 (0:00:02.795)       0:02:40.808 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Thursday 26 March 2026  02:38:32 +0000 (0:00:00.583)       0:02:41.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Thursday 26 March 2026  02:38:35 +0000 (0:00:02.113)       0:02:43.505 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Thursday 26 March 2026  02:38:36 +0000 (0:00:01.906)       0:02:45.411 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Thursday 26 March 2026  02:38:39 +0000 (0:00:02.185)       0:02:47.596 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Thursday 26 March 2026  02:38:41 +0000 (0:00:01.850)       0:02:49.447 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Thursday 26 March 2026  02:38:43 +0000 (0:00:02.833)       0:02:52.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:38:44 +0000 (0:00:00.435)       0:02:52.716 ******** ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 36.85s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 33.11s",
                                "mariadb : Restart MariaDB container ------------------------------------ 12.06s",
                                "mariadb : Check MariaDB service port liveness -------------------------- 10.95s",
                                "mariadb : Running MariaDB bootstrap container -------------------------- 10.18s",
                                "mariadb : Starting first MariaDB container ------------------------------ 7.78s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.01s",
                                "mariadb : Wait for MariaDB service port liveness ------------------------ 4.56s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 3.85s",
                                "mariadb : Copying over config.json files for services ------------------- 3.26s",
                                "mariadb : Ensuring config directories exist ----------------------------- 2.87s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 2.83s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.80s",
                                "mariadb : Check mariadb containers -------------------------------------- 2.68s",
                                "mariadb : Wait for first MariaDB service to sync WSREP ------------------ 2.67s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.58s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 2.27s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.27s",
                                "mariadb : Creating database backup user and setting permissions --------- 2.19s",
                                "mariadb : Creating shard root mysql user -------------------------------- 2.11s",
                                "\u001b[32m2026-03-26 02:38:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639b807-1d34-43fe-916c-01e997ccc45d (rabbitmq) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:38:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b639b807-1d34-43fe-916c-01e997ccc45d (rabbitmq) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:38:51 +0000 (0:00:00.193)       0:00:00.193 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:38:51 +0000 (0:00:00.330)       0:00:00.524 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  02:38:52 +0000 (0:00:00.572)       0:00:01.096 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Thursday 26 March 2026  02:38:52 +0000 (0:00:00.550)       0:00:01.646 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Thursday 26 March 2026  02:38:53 +0000 (0:00:00.918)       0:00:02.565 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Thursday 26 March 2026  02:38:54 +0000 (0:00:00.371)       0:00:02.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Thursday 26 March 2026  02:38:54 +0000 (0:00:00.433)       0:00:03.370 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Thursday 26 March 2026  02:38:55 +0000 (0:00:00.416)       0:00:03.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  02:38:55 +0000 (0:00:00.609)       0:00:04.396 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Thursday 26 March 2026  02:38:56 +0000 (0:00:00.959)       0:00:05.356 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Thursday 26 March 2026  02:38:57 +0000 (0:00:00.781)       0:00:06.137 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Thursday 26 March 2026  02:38:57 +0000 (0:00:00.365)       0:00:06.502 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Thursday 26 March 2026  02:38:58 +0000 (0:00:00.371)       0:00:06.874 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Thursday 26 March 2026  02:38:59 +0000 (0:00:00.826)       0:00:07.700 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Thursday 26 March 2026  02:39:00 +0000 (0:00:01.567)       0:00:09.267 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Thursday 26 March 2026  02:39:01 +0000 (0:00:01.319)       0:00:10.586 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Thursday 26 March 2026  02:39:03 +0000 (0:00:01.714)       0:00:12.301 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Thursday 26 March 2026  02:39:05 +0000 (0:00:01.410)       0:00:13.711 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Thursday 26 March 2026  02:39:06 +0000 (0:00:01.755)       0:00:15.467 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Thursday 26 March 2026  02:39:08 +0000 (0:00:01.430)       0:00:16.898 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  02:39:09 +0000 (0:00:01.454)       0:00:18.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check rabbitmq containers] ************************************",
                                "Thursday 26 March 2026  02:39:10 +0000 (0:00:00.405)       0:00:18.757 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Creating rabbitmq volume] *************************************",
                                "Thursday 26 March 2026  02:39:11 +0000 (0:00:01.184)       0:00:19.942 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Running RabbitMQ bootstrap container] *************************",
                                "Thursday 26 March 2026  02:39:12 +0000 (0:00:00.884)       0:00:20.826 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Thursday 26 March 2026  02:39:19 +0000 (0:00:07.279)       0:00:28.106 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  02:39:20 +0000 (0:00:00.594)       0:00:28.701 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  02:39:20 +0000 (0:00:00.614)       0:00:29.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  02:39:20 +0000 (0:00:00.245)       0:00:29.560 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  02:39:27 +0000 (0:00:06.599)       0:00:36.160 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  02:40:14 +0000 (0:00:46.534)       0:01:22.695 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  02:40:14 +0000 (0:00:00.586)       0:01:23.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  02:40:14 +0000 (0:00:00.254)       0:01:23.536 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  02:40:16 +0000 (0:00:01.691)       0:01:25.228 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  02:40:30 +0000 (0:00:13.585)       0:01:38.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  02:40:30 +0000 (0:00:00.747)       0:01:39.561 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  02:40:31 +0000 (0:00:00.235)       0:01:39.796 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  02:40:32 +0000 (0:00:01.470)       0:01:41.266 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Thursday 26 March 2026  02:40:46 +0000 (0:00:14.155)       0:01:55.422 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Thursday 26 March 2026  02:40:47 +0000 (0:00:00.608)       0:01:56.030 ******** ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35menable_outward_rabbitmq_True\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35moutward_rabbitmq_restart\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq (outward)] *******************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq (outward) services] *************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq (outward) post-configuration] *****************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:40:49 +0000 (0:00:02.403)       0:01:58.434 ******** ",
                                "=============================================================================== ",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------- 74.28s",
                                "rabbitmq : Restart rabbitmq container ----------------------------------- 9.76s",
                                "rabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.28s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 2.40s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 1.95s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 1.76s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 1.71s",
                                "rabbitmq : Copying over config.json files for services ------------------ 1.57s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 1.45s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 1.43s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 1.41s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.32s",
                                "rabbitmq : Check rabbitmq containers ------------------------------------ 1.18s",
                                "rabbitmq : include_tasks ------------------------------------------------ 0.96s",
                                "rabbitmq : Get container facts ------------------------------------------ 0.92s",
                                "rabbitmq : Creating rabbitmq volume ------------------------------------- 0.88s",
                                "rabbitmq : Ensuring config directories exist ---------------------------- 0.83s",
                                "rabbitmq : Get container facts ------------------------------------------ 0.78s",
                                "rabbitmq : Put RabbitMQ node into maintenance mode ---------------------- 0.74s",
                                "rabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 0.61s",
                                "\u001b[32m2026-03-26 02:40:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9d84b916-885e-4b98-a059-63c4a30b9927 (openvswitch) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:40:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9d84b916-885e-4b98-a059-63c4a30b9927 (openvswitch) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:40:57 +0000 (0:00:00.268)       0:00:00.268 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:40:57 +0000 (0:00:00.714)       0:00:00.983 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Thursday 26 March 2026  02:40:58 +0000 (0:00:00.638)       0:00:01.621 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Thursday 26 March 2026  02:40:59 +0000 (0:00:01.168)       0:00:02.790 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Thursday 26 March 2026  02:41:00 +0000 (0:00:01.219)       0:00:04.010 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Thursday 26 March 2026  02:41:02 +0000 (0:00:01.466)       0:00:05.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Thursday 26 March 2026  02:41:03 +0000 (0:00:01.217)       0:00:06.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Thursday 26 March 2026  02:41:04 +0000 (0:00:00.761)       0:00:07.456 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Thursday 26 March 2026  02:41:05 +0000 (0:00:01.473)       0:00:08.929 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Thursday 26 March 2026  02:41:08 +0000 (0:00:02.203)       0:00:11.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Check openvswitch containers] ******************************",
                                "Thursday 26 March 2026  02:41:09 +0000 (0:00:00.982)       0:00:12.115 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:10 +0000 (0:00:01.614)       0:00:13.729 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.330)       0:00:14.060 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.136)       0:00:14.196 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.144)       0:00:14.340 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.153)       0:00:14.494 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.148)       0:00:14.642 ******** ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Thursday 26 March 2026  02:41:11 +0000 (0:00:00.133)       0:00:14.776 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Thursday 26 March 2026  02:41:20 +0000 (0:00:08.407)       0:00:23.183 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Thursday 26 March 2026  02:41:21 +0000 (0:00:01.784)       0:00:24.968 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Thursday 26 March 2026  02:41:29 +0000 (0:00:08.005)       0:00:32.973 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Thursday 26 March 2026  02:41:36 +0000 (0:00:06.330)       0:00:39.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Thursday 26 March 2026  02:41:38 +0000 (0:00:02.424)       0:00:41.728 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Thursday 26 March 2026  02:41:41 +0000 (0:00:03.021)       0:00:44.750 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:41:48 +0000 (0:00:07.104)       0:00:51.855 ******** ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 15.11s",
                                "openvswitch : Restart openvswitch-db-server container ------------------- 8.41s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 6.33s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 3.02s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 2.42s",
                                "openvswitch : Copying over config.json files for services --------------- 2.20s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.78s",
                                "openvswitch : Check openvswitch containers ------------------------------ 1.61s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.47s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.47s",
                                "module-load : Load modules ---------------------------------------------- 1.22s",
                                "module-load : Drop module persistence ----------------------------------- 1.22s",
                                "openvswitch : include_tasks --------------------------------------------- 1.17s",
                                "openvswitch : Flush Handlers -------------------------------------------- 1.05s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.98s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 0.76s",
                                "Group hosts based on Kolla action --------------------------------------- 0.72s",
                                "Group hosts based on enabled services ----------------------------------- 0.64s",
                                "\u001b[32m2026-03-26 02:41:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 13a853e3-ac1d-41a3-984d-fc24ff8bacb1 (ovn) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:41:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 13a853e3-ac1d-41a3-984d-fc24ff8bacb1 (ovn) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  02:41:56 +0000 (0:00:00.203)       0:00:00.203 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  02:41:57 +0000 (0:00:00.774)       0:00:00.978 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Thursday 26 March 2026  02:41:57 +0000 (0:00:00.867)       0:00:01.845 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Thursday 26 March 2026  02:41:59 +0000 (0:00:01.142)       0:00:02.988 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Thursday 26 March 2026  02:42:00 +0000 (0:00:01.188)       0:00:04.177 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Thursday 26 March 2026  02:42:01 +0000 (0:00:01.471)       0:00:05.648 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Thursday 26 March 2026  02:42:02 +0000 (0:00:01.271)       0:00:06.920 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Check ovn-controller containers] ************************",
                                "Thursday 26 March 2026  02:42:04 +0000 (0:00:01.634)       0:00:08.554 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Thursday 26 March 2026  02:42:05 +0000 (0:00:01.357)       0:00:09.912 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Thursday 26 March 2026  02:42:08 +0000 (0:00:02.333)       0:00:12.246 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:26 +0000 (0:00:18.232)       0:00:30.479 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:26 +0000 (0:00:00.260)       0:00:30.739 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:26 +0000 (0:00:00.105)       0:00:30.844 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:26 +0000 (0:00:00.098)       0:00:30.943 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:27 +0000 (0:00:00.071)       0:00:31.014 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  02:42:27 +0000 (0:00:00.073)       0:00:31.087 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Reload systemd config] ***********************",
                                "Thursday 26 March 2026  02:42:27 +0000 (0:00:00.073)       0:00:31.160 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Thursday 26 March 2026  02:42:28 +0000 (0:00:01.581)       0:00:32.741 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  02:43:01 +0000 (0:00:32.834)       0:01:05.575 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  02:43:02 +0000 (0:00:00.730)       0:01:06.306 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Thursday 26 March 2026  02:43:02 +0000 (0:00:00.534)       0:01:06.840 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Thursday 26 March 2026  02:43:03 +0000 (0:00:01.034)       0:01:07.875 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Thursday 26 March 2026  02:43:04 +0000 (0:00:00.338)       0:01:08.214 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Thursday 26 March 2026  02:43:04 +0000 (0:00:00.339)       0:01:08.553 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Thursday 26 March 2026  02:43:04 +0000 (0:00:00.333)       0:01:08.887 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Thursday 26 March 2026  02:43:05 +0000 (0:00:00.568)       0:01:09.456 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Thursday 26 March 2026  02:43:05 +0000 (0:00:00.322)       0:01:09.778 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Thursday 26 March 2026  02:43:06 +0000 (0:00:00.343)       0:01:10.122 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Thursday 26 March 2026  02:43:06 +0000 (0:00:00.338)       0:01:10.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Thursday 26 March 2026  02:43:06 +0000 (0:00:00.308)       0:01:10.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Thursday 26 March 2026  02:43:07 +0000 (0:00:00.547)       0:01:11.316 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Thursday 26 March 2026  02:43:07 +0000 (0:00:00.303)       0:01:11.620 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Thursday 26 March 2026  02:43:07 +0000 (0:00:00.309)       0:01:11.930 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Thursday 26 March 2026  02:43:08 +0000 (0:00:00.292)       0:01:12.222 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Thursday 26 March 2026  02:43:08 +0000 (0:00:00.506)       0:01:12.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Thursday 26 March 2026  02:43:09 +0000 (0:00:00.309)       0:01:13.039 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Thursday 26 March 2026  02:43:09 +0000 (0:00:00.303)       0:01:13.342 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  02:43:09 +0000 (0:00:00.305)       0:01:13.648 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************",
                                "Thursday 26 March 2026  02:43:10 +0000 (0:00:00.763)       0:01:14.411 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************",
                                "Thursday 26 March 2026  02:43:10 +0000 (0:00:00.442)       0:01:14.854 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check NB cluster status] ****************************************",
                                "Thursday 26 March 2026  02:43:11 +0000 (0:00:00.423)       0:01:15.278 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check SB cluster status] ****************************************",
                                "Thursday 26 March 2026  02:43:11 +0000 (0:00:00.334)       0:01:15.613 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***",
                                "Thursday 26 March 2026  02:43:12 +0000 (0:00:00.602)       0:01:16.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***",
                                "Thursday 26 March 2026  02:43:12 +0000 (0:00:00.372)       0:01:16.588 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************",
                                "Thursday 26 March 2026  02:43:12 +0000 (0:00:00.349)       0:01:16.938 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************",
                                "Thursday 26 March 2026  02:43:13 +0000 (0:00:00.330)       0:01:17.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  02:43:13 +0000 (0:00:00.569)       0:01:17.839 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  02:43:15 +0000 (0:00:01.476)       0:01:19.315 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Thursday 26 March 2026  02:43:19 +0000 (0:00:03.899)       0:01:23.215 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:21 +0000 (0:00:02.092)       0:01:25.307 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:21 +0000 (0:00:00.065)       0:01:25.373 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:21 +0000 (0:00:00.289)       0:01:25.662 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Thursday 26 March 2026  02:43:21 +0000 (0:00:00.072)       0:01:25.735 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Thursday 26 March 2026  02:43:29 +0000 (0:00:07.747)       0:01:33.482 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Thursday 26 March 2026  02:43:32 +0000 (0:00:02.620)       0:01:36.103 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Thursday 26 March 2026  02:43:39 +0000 (0:00:07.646)       0:01:43.749 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Thursday 26 March 2026  02:43:39 +0000 (0:00:00.119)       0:01:43.869 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Thursday 26 March 2026  02:43:40 +0000 (0:00:01.058)       0:01:44.928 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Thursday 26 March 2026  02:43:41 +0000 (0:00:00.619)       0:01:45.547 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Thursday 26 March 2026  02:43:42 +0000 (0:00:00.798)       0:01:46.346 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Thursday 26 March 2026  02:43:43 +0000 (0:00:00.638)       0:01:46.984 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Thursday 26 March 2026  02:43:44 +0000 (0:00:01.219)       0:01:48.204 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Unset bootstrap args fact] **************************************",
                                "Thursday 26 March 2026  02:43:45 +0000 (0:00:00.766)       0:01:48.971 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  02:43:45 +0000 (0:00:00.328)       0:01:49.299 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  02:43:46 +0000 (0:00:01.476)       0:01:50.776 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Thursday 26 March 2026  02:43:50 +0000 (0:00:04.001)       0:01:54.777 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:54 +0000 (0:00:03.346)       0:01:58.124 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:54 +0000 (0:00:00.122)       0:01:58.246 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  02:43:54 +0000 (0:00:00.090)       0:01:58.337 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Thursday 26 March 2026  02:43:54 +0000 (0:00:00.099)       0:01:58.436 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Thursday 26 March 2026  02:44:00 +0000 (0:00:06.304)       0:02:04.740 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Thursday 26 March 2026  02:44:07 +0000 (0:00:06.237)       0:02:10.978 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Thursday 26 March 2026  02:44:13 +0000 (0:00:06.207)       0:02:17.186 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Thursday 26 March 2026  02:44:13 +0000 (0:00:00.171)       0:02:17.357 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Thursday 26 March 2026  02:44:14 +0000 (0:00:01.063)       0:02:18.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Thursday 26 March 2026  02:44:15 +0000 (0:00:00.674)       0:02:19.096 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Thursday 26 March 2026  02:44:15 +0000 (0:00:00.822)       0:02:19.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Thursday 26 March 2026  02:44:16 +0000 (0:00:00.641)       0:02:20.561 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Thursday 26 March 2026  02:44:17 +0000 (0:00:01.176)       0:02:21.737 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:44:18 +0000 (0:00:01.000)       0:02:22.738 ******** ",
                                "=============================================================================== ",
                                "ovn-controller : Restart ovn-controller container ---------------------- 32.83s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 18.23s",
                                "ovn-db : Restart ovn-nb-db container ----------------------------------- 14.05s",
                                "ovn-db : Restart ovn-northd container ---------------------------------- 13.85s",
                                "ovn-db : Restart ovn-sb-db container ------------------------------------ 8.86s",
                                "ovn-db : Copying over config.json files for services -------------------- 4.00s",
                                "ovn-db : Copying over config.json files for services -------------------- 3.90s",
                                "ovn-db : Check ovn containers ------------------------------------------- 3.35s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.33s",
                                "ovn-db : Check ovn containers ------------------------------------------- 2.09s",
                                "ovn-controller : Copying over systemd override -------------------------- 1.63s",
                                "ovn-controller : Reload systemd config ---------------------------------- 1.58s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.48s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.48s",
                                "ovn-controller : Copying over config.json files for services ------------ 1.47s",
                                "ovn-controller : Check ovn-controller containers ------------------------ 1.36s",
                                "ovn-controller : Ensuring systemd override directory exists ------------- 1.27s",
                                "ovn-db : Wait for ovn-nb-db --------------------------------------------- 1.22s",
                                "ovn-controller : Ensuring config directories exist ---------------------- 1.19s",
                                "ovn-db : Wait for ovn-nb-db --------------------------------------------- 1.18s",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh",
                                "\u001b[32m2026-03-26 02:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m",
                                "\u001b[32m2026-03-26 02:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 89ea481d-9381-42c6-ad2c-7d8c23c0aae3 (wipe-partitions) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 89ea481d-9381-42c6-ad2c-7d8c23c0aae3 (wipe-partitions) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wipe partitions] *********************************************************",
                                "",
                                "TASK [Find all logical devices owned by UID 167] *******************************",
                                "Thursday 26 March 2026  02:44:36 +0000 (0:00:00.153)       0:00:00.153 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all rook related logical devices] *********************************",
                                "Thursday 26 March 2026  02:44:36 +0000 (0:00:00.590)       0:00:00.744 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Find all logical devices with prefix ceph] *******************************",
                                "Thursday 26 March 2026  02:44:37 +0000 (0:00:00.398)       0:00:01.143 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Remove all ceph related logical devices] *********************************",
                                "Thursday 26 March 2026  02:44:38 +0000 (0:00:00.637)       0:00:01.780 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check device availability] ***********************************************",
                                "Thursday 26 March 2026  02:44:38 +0000 (0:00:00.281)       0:00:02.062 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Wipe partitions with wipefs] *********************************************",
                                "Thursday 26 March 2026  02:44:39 +0000 (0:00:01.215)       0:00:03.278 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Overwrite first 32M with zeros] ******************************************",
                                "Thursday 26 March 2026  02:44:41 +0000 (0:00:01.629)       0:00:04.907 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Reload udev rules] *******************************************************",
                                "Thursday 26 March 2026  02:44:44 +0000 (0:00:03.037)       0:00:07.944 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Request device events from the kernel] ***********************************",
                                "Thursday 26 March 2026  02:44:44 +0000 (0:00:00.649)       0:00:08.594 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:44:45 +0000 (0:00:00.645)       0:00:09.240 ******** ",
                                "=============================================================================== ",
                                "Overwrite first 32M with zeros ------------------------------------------ 3.04s",
                                "Wipe partitions with wipefs --------------------------------------------- 1.63s",
                                "Check device availability ----------------------------------------------- 1.22s",
                                "Reload udev rules ------------------------------------------------------- 0.65s",
                                "Request device events from the kernel ----------------------------------- 0.65s",
                                "Find all logical devices with prefix ceph ------------------------------- 0.64s",
                                "Find all logical devices owned by UID 167 ------------------------------- 0.59s",
                                "Remove all rook related logical devices --------------------------------- 0.40s",
                                "Remove all ceph related logical devices --------------------------------- 0.28s",
                                "\u001b[32m2026-03-26 02:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a5e1532e-72fd-419b-9693-6cea10c48d87 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a5e1532e-72fd-419b-9693-6cea10c48d87 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  02:45:03 +0000 (0:00:00.297)       0:00:00.297 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Thursday 26 March 2026  02:45:04 +0000 (0:00:01.137)       0:00:01.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:45:05 +0000 (0:00:01.370)       0:00:02.805 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Thursday 26 March 2026  02:45:11 +0000 (0:00:06.124)       0:00:08.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:45:12 +0000 (0:00:00.597)       0:00:09.527 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 6.12s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.37s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.14s",
                                "Gather facts for all hosts ---------------------------------------------- 0.60s",
                                "\u001b[32m2026-03-26 02:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b695aa9f-dffc-4083-94dc-70d9d0de54bc (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b695aa9f-dffc-4083-94dc-70d9d0de54bc (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:45:20 +0000 (0:00:00.371)       0:00:00.371 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:45:20 +0000 (0:00:00.281)       0:00:00.653 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:21 +0000 (0:00:00.270)       0:00:00.924 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:21 +0000 (0:00:00.553)       0:00:01.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:21 +0000 (0:00:00.248)       0:00:01.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:22 +0000 (0:00:00.224)       0:00:01.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:22 +0000 (0:00:00.206)       0:00:02.156 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:22 +0000 (0:00:00.215)       0:00:02.372 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:22 +0000 (0:00:00.218)       0:00:02.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:23 +0000 (0:00:00.245)       0:00:02.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:23 +0000 (0:00:00.220)       0:00:03.057 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:23 +0000 (0:00:00.226)       0:00:03.283 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:24 +0000 (0:00:00.708)       0:00:03.992 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:24 +0000 (0:00:00.720)       0:00:04.712 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:25 +0000 (0:00:01.090)       0:00:05.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:26 +0000 (0:00:00.496)       0:00:06.299 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:26 +0000 (0:00:00.382)       0:00:06.682 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:27 +0000 (0:00:00.446)       0:00:07.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:27 +0000 (0:00:00.238)       0:00:07.367 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:27 +0000 (0:00:00.216)       0:00:07.583 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:28 +0000 (0:00:00.228)       0:00:07.812 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:28 +0000 (0:00:00.222)       0:00:08.035 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:28 +0000 (0:00:00.244)       0:00:08.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:29 +0000 (0:00:00.681)       0:00:08.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:29 +0000 (0:00:00.256)       0:00:09.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:29 +0000 (0:00:00.211)       0:00:09.428 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:30 +0000 (0:00:00.783)       0:00:10.211 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:30 +0000 (0:00:00.233)       0:00:10.445 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:30 +0000 (0:00:00.232)       0:00:10.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:31 +0000 (0:00:00.216)       0:00:10.894 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Thursday 26 March 2026  02:45:31 +0000 (0:00:00.232)       0:00:11.126 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Thursday 26 March 2026  02:45:31 +0000 (0:00:00.200)       0:00:11.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Thursday 26 March 2026  02:45:31 +0000 (0:00:00.177)       0:00:11.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Thursday 26 March 2026  02:45:31 +0000 (0:00:00.203)       0:00:11.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Thursday 26 March 2026  02:45:32 +0000 (0:00:00.396)       0:00:12.105 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Thursday 26 March 2026  02:45:32 +0000 (0:00:00.159)       0:00:12.264 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Thursday 26 March 2026  02:45:32 +0000 (0:00:00.208)       0:00:12.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Thursday 26 March 2026  02:45:32 +0000 (0:00:00.161)       0:00:12.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.184)       0:00:12.818 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.175)       0:00:12.993 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.149)       0:00:13.142 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.144)       0:00:13.286 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.151)       0:00:13.438 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.139)       0:00:13.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Thursday 26 March 2026  02:45:33 +0000 (0:00:00.129)       0:00:13.707 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Thursday 26 March 2026  02:45:34 +0000 (0:00:00.391)       0:00:14.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Thursday 26 March 2026  02:45:34 +0000 (0:00:00.154)       0:00:14.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Thursday 26 March 2026  02:45:34 +0000 (0:00:00.134)       0:00:14.388 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Thursday 26 March 2026  02:45:34 +0000 (0:00:00.151)       0:00:14.540 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-e2623153-bc41-510f-8884-ef957bb96082\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Thursday 26 March 2026  02:45:34 +0000 (0:00:00.232)       0:00:14.772 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:45:36 +0000 (0:00:01.998)       0:00:16.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:45:37 +0000 (0:00:00.269)       0:00:17.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:37 +0000 (0:00:00.241)       0:00:17.282 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:38 +0000 (0:00:00.652)       0:00:17.934 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:38 +0000 (0:00:00.239)       0:00:18.174 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:38 +0000 (0:00:00.243)       0:00:18.417 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:38 +0000 (0:00:00.228)       0:00:18.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:39 +0000 (0:00:00.221)       0:00:18.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:39 +0000 (0:00:00.233)       0:00:19.101 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:39 +0000 (0:00:00.210)       0:00:19.312 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:39 +0000 (0:00:00.212)       0:00:19.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:39 +0000 (0:00:00.212)       0:00:19.738 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:40 +0000 (0:00:00.700)       0:00:20.438 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:41 +0000 (0:00:00.735)       0:00:21.173 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:42 +0000 (0:00:00.966)       0:00:22.140 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:42 +0000 (0:00:00.480)       0:00:22.620 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:43 +0000 (0:00:00.368)       0:00:22.989 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:43 +0000 (0:00:00.424)       0:00:23.414 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:43 +0000 (0:00:00.219)       0:00:23.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:44 +0000 (0:00:00.208)       0:00:23.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:44 +0000 (0:00:00.206)       0:00:24.048 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:44 +0000 (0:00:00.230)       0:00:24.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:44 +0000 (0:00:00.220)       0:00:24.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:44 +0000 (0:00:00.218)       0:00:24.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:45 +0000 (0:00:00.212)       0:00:24.931 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:45 +0000 (0:00:00.726)       0:00:25.658 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:46 +0000 (0:00:00.727)       0:00:26.386 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:46 +0000 (0:00:00.221)       0:00:26.607 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.224)       0:00:26.832 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.270)       0:00:27.103 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.211)       0:00:27.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.179)       0:00:27.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.135)       0:00:27.629 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Thursday 26 March 2026  02:45:47 +0000 (0:00:00.141)       0:00:27.771 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Thursday 26 March 2026  02:45:48 +0000 (0:00:00.145)       0:00:27.916 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Thursday 26 March 2026  02:45:48 +0000 (0:00:00.130)       0:00:28.046 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Thursday 26 March 2026  02:45:48 +0000 (0:00:00.199)       0:00:28.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Thursday 26 March 2026  02:45:48 +0000 (0:00:00.397)       0:00:28.643 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Thursday 26 March 2026  02:45:48 +0000 (0:00:00.157)       0:00:28.800 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.152)       0:00:28.953 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.138)       0:00:29.091 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.149)       0:00:29.241 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.143)       0:00:29.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.141)       0:00:29.526 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Thursday 26 March 2026  02:45:49 +0000 (0:00:00.153)       0:00:29.680 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Thursday 26 March 2026  02:45:50 +0000 (0:00:00.147)       0:00:29.827 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Thursday 26 March 2026  02:45:50 +0000 (0:00:00.132)       0:00:29.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Thursday 26 March 2026  02:45:50 +0000 (0:00:00.140)       0:00:30.100 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Thursday 26 March 2026  02:45:50 +0000 (0:00:00.163)       0:00:30.264 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-a652979e-9f40-503a-bbc8-6de5e605991e\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Thursday 26 March 2026  02:45:50 +0000 (0:00:00.467)       0:00:30.731 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:45:52 +0000 (0:00:01.224)       0:00:31.956 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:45:52 +0000 (0:00:00.289)       0:00:32.246 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:52 +0000 (0:00:00.275)       0:00:32.521 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:53 +0000 (0:00:00.401)       0:00:32.923 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:53 +0000 (0:00:00.203)       0:00:33.126 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:53 +0000 (0:00:00.198)       0:00:33.325 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:53 +0000 (0:00:00.234)       0:00:33.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:54 +0000 (0:00:00.703)       0:00:34.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:54 +0000 (0:00:00.228)       0:00:34.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:54 +0000 (0:00:00.219)       0:00:34.710 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:55 +0000 (0:00:00.226)       0:00:34.937 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:55 +0000 (0:00:00.205)       0:00:35.142 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:55 +0000 (0:00:00.444)       0:00:35.587 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:56 +0000 (0:00:00.502)       0:00:36.089 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:56 +0000 (0:00:00.453)       0:00:36.543 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:45:57 +0000 (0:00:00.511)       0:00:37.055 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:57 +0000 (0:00:00.407)       0:00:37.462 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:58 +0000 (0:00:00.682)       0:00:38.145 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:58 +0000 (0:00:00.204)       0:00:38.349 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:58 +0000 (0:00:00.228)       0:00:38.578 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:58 +0000 (0:00:00.211)       0:00:38.789 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:59 +0000 (0:00:00.253)       0:00:39.043 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:59 +0000 (0:00:00.207)       0:00:39.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:59 +0000 (0:00:00.227)       0:00:39.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:45:59 +0000 (0:00:00.219)       0:00:39.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:46:00 +0000 (0:00:00.236)       0:00:39.934 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:46:01 +0000 (0:00:00.952)       0:00:40.887 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:46:01 +0000 (0:00:00.188)       0:00:41.075 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:46:01 +0000 (0:00:00.194)       0:00:41.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:46:02 +0000 (0:00:00.762)       0:00:42.032 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Thursday 26 March 2026  02:46:02 +0000 (0:00:00.220)       0:00:42.252 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Thursday 26 March 2026  02:46:02 +0000 (0:00:00.197)       0:00:42.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Thursday 26 March 2026  02:46:02 +0000 (0:00:00.141)       0:00:42.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Thursday 26 March 2026  02:46:02 +0000 (0:00:00.148)       0:00:42.740 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.133)       0:00:42.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.168)       0:00:43.043 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.188)       0:00:43.231 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.152)       0:00:43.383 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.179)       0:00:43.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Thursday 26 March 2026  02:46:03 +0000 (0:00:00.155)       0:00:43.717 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Thursday 26 March 2026  02:46:04 +0000 (0:00:00.170)       0:00:43.888 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Thursday 26 March 2026  02:46:04 +0000 (0:00:00.393)       0:00:44.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Thursday 26 March 2026  02:46:04 +0000 (0:00:00.153)       0:00:44.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Thursday 26 March 2026  02:46:04 +0000 (0:00:00.152)       0:00:44.588 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Thursday 26 March 2026  02:46:04 +0000 (0:00:00.134)       0:00:44.723 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Thursday 26 March 2026  02:46:05 +0000 (0:00:00.156)       0:00:44.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Thursday 26 March 2026  02:46:05 +0000 (0:00:00.145)       0:00:45.025 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Thursday 26 March 2026  02:46:05 +0000 (0:00:00.145)       0:00:45.171 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Thursday 26 March 2026  02:46:05 +0000 (0:00:00.139)       0:00:45.310 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Thursday 26 March 2026  02:46:05 +0000 (0:00:00.222)       0:00:45.533 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:46:06 +0000 (0:00:01.078)       0:00:46.612 ******** ",
                                "=============================================================================== ",
                                "Write configuration file ------------------------------------------------ 4.30s",
                                "Add known links to the list of available block devices ------------------ 1.61s",
                                "Add known partitions to the list of available block devices ------------- 1.55s",
                                "Add known links to the list of available block devices ------------------ 1.09s",
                                "Add known links to the list of available block devices ------------------ 0.97s",
                                "Add known partitions to the list of available block devices ------------- 0.95s",
                                "Print configuration data ------------------------------------------------ 0.92s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.84s",
                                "Get initial list of available block devices ----------------------------- 0.79s",
                                "Add known partitions to the list of available block devices ------------- 0.78s",
                                "Add known partitions to the list of available block devices ------------- 0.76s",
                                "Add known links to the list of available block devices ------------------ 0.74s",
                                "Add known partitions to the list of available block devices ------------- 0.73s",
                                "Add known partitions to the list of available block devices ------------- 0.73s",
                                "Add known links to the list of available block devices ------------------ 0.72s",
                                "Generate lvm_volumes structure (block + db) ----------------------------- 0.71s",
                                "Add known links to the list of available block devices ------------------ 0.71s",
                                "Add known links to the list of available block devices ------------------ 0.70s",
                                "Add known links to the list of available block devices ------------------ 0.70s",
                                "Print ceph_osd_devices -------------------------------------------------- 0.70s",
                                "\u001b[32m2026-03-26 02:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bfb64ee4-2646-4665-9bf0-88d97a04f762 (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-03-26 02:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "[master 7661ed7] 2026-03-26-02-47",
                                " 1 file changed, 30 insertions(+), 9 deletions(-)",
                                "\u001b[32m2026-03-26 02:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 84ec70a8-97fe-4cd7-9d64-7d2214c4453b (ceph-create-lvm-devices) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 84ec70a8-97fe-4cd7-9d64-7d2214c4453b (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:47:08 +0000 (0:00:00.339)       0:00:00.339 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:47:08 +0000 (0:00:00.268)       0:00:00.608 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:09 +0000 (0:00:00.308)       0:00:00.916 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:09 +0000 (0:00:00.593)       0:00:01.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:10 +0000 (0:00:00.251)       0:00:01.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:10 +0000 (0:00:00.234)       0:00:01.997 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:10 +0000 (0:00:00.243)       0:00:02.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:10 +0000 (0:00:00.245)       0:00:02.486 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:10 +0000 (0:00:00.225)       0:00:02.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:11 +0000 (0:00:00.236)       0:00:02.948 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:11 +0000 (0:00:00.229)       0:00:03.177 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:11 +0000 (0:00:00.225)       0:00:03.403 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:12 +0000 (0:00:00.713)       0:00:04.116 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:13 +0000 (0:00:00.717)       0:00:04.834 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:14 +0000 (0:00:00.979)       0:00:05.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:14 +0000 (0:00:00.477)       0:00:06.290 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:14 +0000 (0:00:00.334)       0:00:06.625 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:15 +0000 (0:00:00.461)       0:00:07.087 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:15 +0000 (0:00:00.231)       0:00:07.318 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:15 +0000 (0:00:00.218)       0:00:07.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:16 +0000 (0:00:00.212)       0:00:07.750 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:16 +0000 (0:00:00.214)       0:00:07.964 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:16 +0000 (0:00:00.203)       0:00:08.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:17 +0000 (0:00:00.670)       0:00:08.839 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:17 +0000 (0:00:00.209)       0:00:09.048 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:17 +0000 (0:00:00.211)       0:00:09.259 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:18 +0000 (0:00:00.731)       0:00:09.991 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:18 +0000 (0:00:00.247)       0:00:10.238 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:18 +0000 (0:00:00.241)       0:00:10.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:18 +0000 (0:00:00.229)       0:00:10.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Thursday 26 March 2026  02:47:19 +0000 (0:00:00.238)       0:00:10.948 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Thursday 26 March 2026  02:47:19 +0000 (0:00:00.154)       0:00:11.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'e2623153-bc41-510f-8884-ef957bb96082'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Thursday 26 March 2026  02:47:19 +0000 (0:00:00.217)       0:00:11.320 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Thursday 26 March 2026  02:47:21 +0000 (0:00:02.059)       0:00:13.379 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Thursday 26 March 2026  02:47:22 +0000 (0:00:00.384)       0:00:13.764 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Thursday 26 March 2026  02:47:23 +0000 (0:00:01.510)       0:00:15.275 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Thursday 26 March 2026  02:47:23 +0000 (0:00:00.154)       0:00:15.430 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Thursday 26 March 2026  02:47:23 +0000 (0:00:00.157)       0:00:15.587 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.175)       0:00:15.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.146)       0:00:15.909 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.161)       0:00:16.071 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.141)       0:00:16.212 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.181)       0:00:16.393 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.156)       0:00:16.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Thursday 26 March 2026  02:47:24 +0000 (0:00:00.168)       0:00:16.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Thursday 26 March 2026  02:47:25 +0000 (0:00:00.406)       0:00:17.125 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Thursday 26 March 2026  02:47:25 +0000 (0:00:00.238)       0:00:17.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Thursday 26 March 2026  02:47:25 +0000 (0:00:00.143)       0:00:17.507 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Thursday 26 March 2026  02:47:25 +0000 (0:00:00.150)       0:00:17.657 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Thursday 26 March 2026  02:47:26 +0000 (0:00:00.151)       0:00:17.808 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Thursday 26 March 2026  02:47:26 +0000 (0:00:00.156)       0:00:17.965 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Thursday 26 March 2026  02:47:26 +0000 (0:00:00.154)       0:00:18.119 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Thursday 26 March 2026  02:47:26 +0000 (0:00:00.178)       0:00:18.298 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Thursday 26 March 2026  02:47:27 +0000 (0:00:00.701)       0:00:18.999 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Thursday 26 March 2026  02:47:27 +0000 (0:00:00.503)       0:00:19.502 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Thursday 26 March 2026  02:47:28 +0000 (0:00:00.523)       0:00:20.026 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Thursday 26 March 2026  02:47:28 +0000 (0:00:00.389)       0:00:20.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Thursday 26 March 2026  02:47:28 +0000 (0:00:00.120)       0:00:20.536 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Thursday 26 March 2026  02:47:28 +0000 (0:00:00.126)       0:00:20.663 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.147)       0:00:20.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.138)       0:00:20.949 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.143)       0:00:21.093 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.154)       0:00:21.247 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.151)       0:00:21.399 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.155)       0:00:21.554 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Thursday 26 March 2026  02:47:29 +0000 (0:00:00.154)       0:00:21.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Thursday 26 March 2026  02:47:30 +0000 (0:00:00.144)       0:00:21.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Thursday 26 March 2026  02:47:30 +0000 (0:00:00.370)       0:00:22.224 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Thursday 26 March 2026  02:47:30 +0000 (0:00:00.149)       0:00:22.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:47:30 +0000 (0:00:00.162)       0:00:22.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Thursday 26 March 2026  02:47:30 +0000 (0:00:00.156)       0:00:22.692 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.149)       0:00:22.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.150)       0:00:22.992 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.142)       0:00:23.134 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.170)       0:00:23.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.167)       0:00:23.472 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Thursday 26 March 2026  02:47:31 +0000 (0:00:00.166)       0:00:23.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Thursday 26 March 2026  02:47:32 +0000 (0:00:00.166)       0:00:23.806 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Thursday 26 March 2026  02:47:32 +0000 (0:00:00.415)       0:00:24.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Thursday 26 March 2026  02:47:32 +0000 (0:00:00.171)       0:00:24.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Thursday 26 March 2026  02:47:32 +0000 (0:00:00.180)       0:00:24.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:47:33 +0000 (0:00:00.184)       0:00:24.757 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:47:33 +0000 (0:00:00.531)       0:00:25.289 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Thursday 26 March 2026  02:47:34 +0000 (0:00:00.530)       0:00:25.819 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Thursday 26 March 2026  02:47:34 +0000 (0:00:00.165)       0:00:25.985 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'vg_name': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'vg_name': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Thursday 26 March 2026  02:47:34 +0000 (0:00:00.211)       0:00:26.197 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Thursday 26 March 2026  02:47:34 +0000 (0:00:00.180)       0:00:26.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Thursday 26 March 2026  02:47:34 +0000 (0:00:00.186)       0:00:26.564 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Thursday 26 March 2026  02:47:35 +0000 (0:00:00.190)       0:00:26.754 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-e2623153-bc41-510f-8884-ef957bb96082\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-e2623153-bc41-510f-8884-ef957bb96082\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:47:35 +0000 (0:00:00.577)       0:00:27.332 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:47:35 +0000 (0:00:00.281)       0:00:27.613 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:36 +0000 (0:00:00.254)       0:00:27.868 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:36 +0000 (0:00:00.487)       0:00:28.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:36 +0000 (0:00:00.219)       0:00:28.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:37 +0000 (0:00:00.232)       0:00:28.807 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:37 +0000 (0:00:00.214)       0:00:29.022 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:37 +0000 (0:00:00.206)       0:00:29.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:37 +0000 (0:00:00.229)       0:00:29.457 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:38 +0000 (0:00:00.689)       0:00:30.147 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:38 +0000 (0:00:00.240)       0:00:30.387 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:38 +0000 (0:00:00.218)       0:00:30.606 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:39 +0000 (0:00:00.445)       0:00:31.051 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:39 +0000 (0:00:00.482)       0:00:31.534 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:40 +0000 (0:00:00.486)       0:00:32.021 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:47:40 +0000 (0:00:00.503)       0:00:32.524 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:41 +0000 (0:00:00.376)       0:00:32.900 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:41 +0000 (0:00:00.439)       0:00:33.340 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:41 +0000 (0:00:00.210)       0:00:33.551 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:42 +0000 (0:00:00.248)       0:00:33.799 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:42 +0000 (0:00:00.706)       0:00:34.506 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:42 +0000 (0:00:00.221)       0:00:34.727 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:43 +0000 (0:00:00.239)       0:00:34.966 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:43 +0000 (0:00:00.215)       0:00:35.182 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:43 +0000 (0:00:00.229)       0:00:35.412 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:43 +0000 (0:00:00.230)       0:00:35.642 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:44 +0000 (0:00:00.728)       0:00:36.371 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:44 +0000 (0:00:00.236)       0:00:36.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:45 +0000 (0:00:00.222)       0:00:36.831 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:47:45 +0000 (0:00:00.241)       0:00:37.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Thursday 26 March 2026  02:47:45 +0000 (0:00:00.219)       0:00:37.292 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Thursday 26 March 2026  02:47:45 +0000 (0:00:00.384)       0:00:37.677 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a652979e-9f40-503a-bbc8-6de5e605991e'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'b5eee7c3-8883-5bbe-be5a-75726e822543'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Thursday 26 March 2026  02:47:46 +0000 (0:00:00.206)       0:00:37.883 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Thursday 26 March 2026  02:47:47 +0000 (0:00:01.739)       0:00:39.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Thursday 26 March 2026  02:47:48 +0000 (0:00:00.244)       0:00:39.867 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Thursday 26 March 2026  02:47:49 +0000 (0:00:01.289)       0:00:41.156 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Thursday 26 March 2026  02:47:49 +0000 (0:00:00.187)       0:00:41.344 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Thursday 26 March 2026  02:47:49 +0000 (0:00:00.157)       0:00:41.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Thursday 26 March 2026  02:47:49 +0000 (0:00:00.183)       0:00:41.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Thursday 26 March 2026  02:47:50 +0000 (0:00:00.149)       0:00:41.835 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Thursday 26 March 2026  02:47:50 +0000 (0:00:00.159)       0:00:41.995 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Thursday 26 March 2026  02:47:50 +0000 (0:00:00.169)       0:00:42.164 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Thursday 26 March 2026  02:47:50 +0000 (0:00:00.165)       0:00:42.330 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Thursday 26 March 2026  02:47:50 +0000 (0:00:00.197)       0:00:42.528 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Thursday 26 March 2026  02:47:51 +0000 (0:00:00.448)       0:00:42.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Thursday 26 March 2026  02:47:51 +0000 (0:00:00.161)       0:00:43.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Thursday 26 March 2026  02:47:51 +0000 (0:00:00.167)       0:00:43.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Thursday 26 March 2026  02:47:51 +0000 (0:00:00.172)       0:00:43.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Thursday 26 March 2026  02:47:51 +0000 (0:00:00.142)       0:00:43.620 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Thursday 26 March 2026  02:47:52 +0000 (0:00:00.155)       0:00:43.776 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Thursday 26 March 2026  02:47:52 +0000 (0:00:00.141)       0:00:43.917 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Thursday 26 March 2026  02:47:52 +0000 (0:00:00.147)       0:00:44.065 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Thursday 26 March 2026  02:47:52 +0000 (0:00:00.157)       0:00:44.223 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Thursday 26 March 2026  02:47:53 +0000 (0:00:00.520)       0:00:44.743 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Thursday 26 March 2026  02:47:53 +0000 (0:00:00.497)       0:00:45.241 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Thursday 26 March 2026  02:47:53 +0000 (0:00:00.496)       0:00:45.738 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Thursday 26 March 2026  02:47:54 +0000 (0:00:00.418)       0:00:46.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Thursday 26 March 2026  02:47:54 +0000 (0:00:00.124)       0:00:46.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Thursday 26 March 2026  02:47:54 +0000 (0:00:00.133)       0:00:46.415 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Thursday 26 March 2026  02:47:54 +0000 (0:00:00.159)       0:00:46.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Thursday 26 March 2026  02:47:54 +0000 (0:00:00.154)       0:00:46.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.144)       0:00:46.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.171)       0:00:47.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.145)       0:00:47.191 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.143)       0:00:47.334 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.150)       0:00:47.484 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Thursday 26 March 2026  02:47:55 +0000 (0:00:00.139)       0:00:47.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Thursday 26 March 2026  02:47:56 +0000 (0:00:00.156)       0:00:47.781 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Thursday 26 March 2026  02:47:56 +0000 (0:00:00.375)       0:00:48.156 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:47:56 +0000 (0:00:00.148)       0:00:48.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Thursday 26 March 2026  02:47:56 +0000 (0:00:00.207)       0:00:48.513 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Thursday 26 March 2026  02:47:56 +0000 (0:00:00.156)       0:00:48.669 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.171)       0:00:48.841 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.150)       0:00:48.991 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.213)       0:00:49.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.183)       0:00:49.388 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.160)       0:00:49.549 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Thursday 26 March 2026  02:47:57 +0000 (0:00:00.153)       0:00:49.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Thursday 26 March 2026  02:47:58 +0000 (0:00:00.174)       0:00:49.877 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Thursday 26 March 2026  02:47:58 +0000 (0:00:00.156)       0:00:50.034 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Thursday 26 March 2026  02:47:58 +0000 (0:00:00.413)       0:00:50.448 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:47:58 +0000 (0:00:00.182)       0:00:50.631 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:47:59 +0000 (0:00:00.498)       0:00:51.129 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Thursday 26 March 2026  02:47:59 +0000 (0:00:00.563)       0:00:51.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Thursday 26 March 2026  02:48:00 +0000 (0:00:00.177)       0:00:51.870 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'vg_name': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'vg_name': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Thursday 26 March 2026  02:48:00 +0000 (0:00:00.169)       0:00:52.039 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Thursday 26 March 2026  02:48:00 +0000 (0:00:00.166)       0:00:52.206 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Thursday 26 March 2026  02:48:00 +0000 (0:00:00.164)       0:00:52.370 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Thursday 26 March 2026  02:48:00 +0000 (0:00:00.169)       0:00:52.539 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-a652979e-9f40-503a-bbc8-6de5e605991e\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-a652979e-9f40-503a-bbc8-6de5e605991e\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-b5eee7c3-8883-5bbe-be5a-75726e822543\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  02:48:01 +0000 (0:00:00.319)       0:00:52.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Thursday 26 March 2026  02:48:01 +0000 (0:00:00.763)       0:00:53.622 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:02 +0000 (0:00:00.261)       0:00:53.883 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:02 +0000 (0:00:00.428)       0:00:54.311 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:02 +0000 (0:00:00.209)       0:00:54.521 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:03 +0000 (0:00:00.238)       0:00:54.759 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:03 +0000 (0:00:00.208)       0:00:54.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:03 +0000 (0:00:00.231)       0:00:55.199 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:03 +0000 (0:00:00.225)       0:00:55.425 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:03 +0000 (0:00:00.221)       0:00:55.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:04 +0000 (0:00:00.238)       0:00:55.885 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:04 +0000 (0:00:00.205)       0:00:56.090 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:05 +0000 (0:00:00.934)       0:00:57.024 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:05 +0000 (0:00:00.451)       0:00:57.476 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:06 +0000 (0:00:00.474)       0:00:57.951 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Thursday 26 March 2026  02:48:06 +0000 (0:00:00.494)       0:00:58.445 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:07 +0000 (0:00:00.362)       0:00:58.808 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:07 +0000 (0:00:00.454)       0:00:59.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:07 +0000 (0:00:00.222)       0:00:59.485 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:07 +0000 (0:00:00.210)       0:00:59.695 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:08 +0000 (0:00:00.223)       0:00:59.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:08 +0000 (0:00:00.241)       0:01:00.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:09 +0000 (0:00:00.699)       0:01:00.860 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:09 +0000 (0:00:00.234)       0:01:01.094 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:09 +0000 (0:00:00.219)       0:01:01.314 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:09 +0000 (0:00:00.269)       0:01:01.583 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:10 +0000 (0:00:00.720)       0:01:02.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:10 +0000 (0:00:00.227)       0:01:02.531 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:11 +0000 (0:00:00.217)       0:01:02.749 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Thursday 26 March 2026  02:48:11 +0000 (0:00:00.233)       0:01:02.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Thursday 26 March 2026  02:48:11 +0000 (0:00:00.221)       0:01:03.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Thursday 26 March 2026  02:48:11 +0000 (0:00:00.145)       0:01:03.350 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '83c4def8-4703-5f7c-9549-7666ff9f2b66'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Thursday 26 March 2026  02:48:11 +0000 (0:00:00.178)       0:01:03.529 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Thursday 26 March 2026  02:48:13 +0000 (0:00:01.905)       0:01:05.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Thursday 26 March 2026  02:48:14 +0000 (0:00:00.418)       0:01:05.852 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Thursday 26 March 2026  02:48:15 +0000 (0:00:01.377)       0:01:07.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Thursday 26 March 2026  02:48:15 +0000 (0:00:00.180)       0:01:07.410 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Thursday 26 March 2026  02:48:15 +0000 (0:00:00.166)       0:01:07.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.202)       0:01:07.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.166)       0:01:07.945 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.163)       0:01:08.109 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.153)       0:01:08.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.161)       0:01:08.424 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Thursday 26 March 2026  02:48:16 +0000 (0:00:00.153)       0:01:08.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Thursday 26 March 2026  02:48:17 +0000 (0:00:00.170)       0:01:08.747 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Thursday 26 March 2026  02:48:17 +0000 (0:00:00.183)       0:01:08.931 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Thursday 26 March 2026  02:48:17 +0000 (0:00:00.416)       0:01:09.348 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Thursday 26 March 2026  02:48:17 +0000 (0:00:00.155)       0:01:09.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Thursday 26 March 2026  02:48:17 +0000 (0:00:00.145)       0:01:09.649 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Thursday 26 March 2026  02:48:18 +0000 (0:00:00.155)       0:01:09.804 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Thursday 26 March 2026  02:48:18 +0000 (0:00:00.224)       0:01:10.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Thursday 26 March 2026  02:48:18 +0000 (0:00:00.176)       0:01:10.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Thursday 26 March 2026  02:48:18 +0000 (0:00:00.148)       0:01:10.354 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Thursday 26 March 2026  02:48:19 +0000 (0:00:00.528)       0:01:10.883 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Thursday 26 March 2026  02:48:19 +0000 (0:00:00.541)       0:01:11.424 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Thursday 26 March 2026  02:48:20 +0000 (0:00:00.518)       0:01:11.943 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Thursday 26 March 2026  02:48:20 +0000 (0:00:00.157)       0:01:12.100 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Thursday 26 March 2026  02:48:20 +0000 (0:00:00.131)       0:01:12.232 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Thursday 26 March 2026  02:48:20 +0000 (0:00:00.384)       0:01:12.616 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.164)       0:01:12.780 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.173)       0:01:12.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.156)       0:01:13.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.138)       0:01:13.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.171)       0:01:13.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.166)       0:01:13.587 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Thursday 26 March 2026  02:48:21 +0000 (0:00:00.127)       0:01:13.714 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Thursday 26 March 2026  02:48:22 +0000 (0:00:00.152)       0:01:13.867 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Thursday 26 March 2026  02:48:22 +0000 (0:00:00.136)       0:01:14.003 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Thursday 26 March 2026  02:48:22 +0000 (0:00:00.151)       0:01:14.154 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:48:22 +0000 (0:00:00.157)       0:01:14.312 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Thursday 26 March 2026  02:48:22 +0000 (0:00:00.390)       0:01:14.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.158)       0:01:14.861 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.160)       0:01:15.021 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.153)       0:01:15.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.174)       0:01:15.349 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.184)       0:01:15.534 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Thursday 26 March 2026  02:48:23 +0000 (0:00:00.172)       0:01:15.706 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Thursday 26 March 2026  02:48:24 +0000 (0:00:00.191)       0:01:15.898 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Thursday 26 March 2026  02:48:24 +0000 (0:00:00.179)       0:01:16.077 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Thursday 26 March 2026  02:48:24 +0000 (0:00:00.162)       0:01:16.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Thursday 26 March 2026  02:48:24 +0000 (0:00:00.199)       0:01:16.440 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:48:24 +0000 (0:00:00.167)       0:01:16.607 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Thursday 26 March 2026  02:48:25 +0000 (0:00:00.744)       0:01:17.352 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Thursday 26 March 2026  02:48:26 +0000 (0:00:00.515)       0:01:17.868 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Thursday 26 March 2026  02:48:26 +0000 (0:00:00.185)       0:01:18.054 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'vg_name': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'vg_name': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Thursday 26 March 2026  02:48:26 +0000 (0:00:00.185)       0:01:18.239 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Thursday 26 March 2026  02:48:26 +0000 (0:00:00.187)       0:01:18.427 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Thursday 26 March 2026  02:48:26 +0000 (0:00:00.175)       0:01:18.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Thursday 26 March 2026  02:48:27 +0000 (0:00:00.184)       0:01:18.786 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:48:27 +0000 (0:00:00.154)       0:01:18.940 ******** ",
                                "=============================================================================== ",
                                "Create block VGs -------------------------------------------------------- 5.70s",
                                "Create block LVs -------------------------------------------------------- 4.18s",
                                "Get list of Ceph LVs with associated VGs -------------------------------- 1.77s",
                                "Gather DB VGs with total and available size in bytes -------------------- 1.75s",
                                "Get list of Ceph PVs with associated VGs -------------------------------- 1.61s",
                                "Gather WAL VGs with total and available size in bytes ------------------- 1.54s",
                                "Gather DB+WAL VGs with total and available size in bytes ---------------- 1.54s",
                                "Add known links to the list of available block devices ------------------ 1.51s",
                                "Add known partitions to the list of available block devices ------------- 1.36s",
                                "Get extra vars for Ceph configuration ----------------------------------- 1.31s",
                                "Print LVM report data --------------------------------------------------- 1.05s",
                                "Print 'Create block VGs' ------------------------------------------------ 1.05s",
                                "Add known links to the list of available block devices ------------------ 0.98s",
                                "Combine JSON from _db/wal/db_wal_vgs_cmd_output ------------------------- 0.97s",
                                "Add known links to the list of available block devices ------------------ 0.93s",
                                "Get initial list of available block devices ----------------------------- 0.82s",
                                "Count OSDs put on ceph_db_wal_devices defined in lvm_volumes ------------ 0.82s",
                                "Create DB LVs for ceph_db_wal_devices ----------------------------------- 0.79s",
                                "Count OSDs put on ceph_db_devices defined in lvm_volumes ---------------- 0.79s",
                                "Create WAL LVs for ceph_db_wal_devices ---------------------------------- 0.77s",
                                "\u001b[32m2026-03-26 02:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1dc7d21e-cb25-4027-b145-729ca3421def (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1dc7d21e-cb25-4027-b145-729ca3421def (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  02:48:44 +0000 (0:00:00.280)       0:00:00.280 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Thursday 26 March 2026  02:48:45 +0000 (0:00:01.150)       0:00:01.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  02:48:47 +0000 (0:00:01.369)       0:00:02.800 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Thursday 26 March 2026  02:48:52 +0000 (0:00:04.942)       0:00:07.742 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  02:48:52 +0000 (0:00:00.595)       0:00:08.338 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.94s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.37s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.15s",
                                "Gather facts for all hosts ---------------------------------------------- 0.60s",
                                "\u001b[32m2026-03-26 02:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0afd95ab-3afe-4cf0-a623-ba200f82b57e (ceph) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 02:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0afd95ab-3afe-4cf0-a623-ba200f82b57e (ceph) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Prepare deployment of Ceph services] *************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  02:49:01 +0000 (0:00:00.861)       0:00:00.861 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  02:49:02 +0000 (0:00:01.228)       0:00:02.090 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  02:49:03 +0000 (0:00:01.368)       0:00:03.459 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  02:49:04 +0000 (0:00:01.039)       0:00:04.498 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  02:49:06 +0000 (0:00:01.044)       0:00:05.542 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  02:49:07 +0000 (0:00:01.090)       0:00:06.633 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  02:49:07 +0000 (0:00:00.670)       0:00:07.304 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  02:49:08 +0000 (0:00:00.857)       0:00:08.161 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  02:49:09 +0000 (0:00:00.633)       0:00:08.795 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  02:49:10 +0000 (0:00:00.826)       0:00:09.621 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  02:49:10 +0000 (0:00:00.714)       0:00:10.336 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  02:49:11 +0000 (0:00:00.712)       0:00:11.048 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  02:49:14 +0000 (0:00:02.505)       0:00:13.554 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  02:49:14 +0000 (0:00:00.438)       0:00:13.992 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  02:49:15 +0000 (0:00:00.653)       0:00:14.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  02:49:15 +0000 (0:00:00.169)       0:00:14.816 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 02:49:12.428284', 'end': '2026-03-26 02:49:12.465831', 'delta': '0:00:00.037547', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 02:49:13.089373', 'end': '2026-03-26 02:49:13.140350', 'delta': '0:00:00.050977', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 02:49:13.625707', 'end': '2026-03-26 02:49:13.679965', 'delta': '0:00:00.054258', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  02:49:15 +0000 (0:00:00.270)       0:00:15.086 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  02:49:16 +0000 (0:00:00.749)       0:00:15.836 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  02:49:17 +0000 (0:00:01.088)       0:00:16.925 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  02:49:18 +0000 (0:00:00.632)       0:00:17.558 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  02:49:19 +0000 (0:00:01.181)       0:00:18.739 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  02:49:19 +0000 (0:00:00.710)       0:00:19.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  02:49:20 +0000 (0:00:00.145)       0:00:19.595 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  02:49:20 +0000 (0:00:00.234)       0:00:19.829 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  02:49:21 +0000 (0:00:00.828)       0:00:20.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  02:49:21 +0000 (0:00:00.666)       0:00:21.325 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  02:49:22 +0000 (0:00:00.836)       0:00:22.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  02:49:23 +0000 (0:00:00.625)       0:00:22.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  02:49:24 +0000 (0:00:00.880)       0:00:23.668 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  02:49:24 +0000 (0:00:00.647)       0:00:24.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  02:49:25 +0000 (0:00:00.917)       0:00:25.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-12-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-16-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  02:49:26 +0000 (0:00:01.059)       0:00:26.292 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-12-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-16-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  02:49:28 +0000 (0:00:01.520)       0:00:27.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  02:49:29 +0000 (0:00:00.948)       0:00:28.761 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  02:49:30 +0000 (0:00:00.885)       0:00:29.647 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  02:49:31 +0000 (0:00:00.928)       0:00:30.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  02:49:31 +0000 (0:00:00.926)       0:00:31.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  02:49:32 +0000 (0:00:00.664)       0:00:32.166 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  02:49:33 +0000 (0:00:00.925)       0:00:33.092 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  02:49:35 +0000 (0:00:01.820)       0:00:34.912 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  02:49:36 +0000 (0:00:01.050)       0:00:35.963 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  02:49:37 +0000 (0:00:01.128)       0:00:37.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  02:49:37 +0000 (0:00:00.347)       0:00:37.439 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  02:49:38 +0000 (0:00:00.379)       0:00:37.819 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  02:49:38 +0000 (0:00:00.534)       0:00:38.354 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  02:49:39 +0000 (0:00:00.460)       0:00:38.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  02:49:39 +0000 (0:00:00.387)       0:00:39.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  02:49:40 +0000 (0:00:00.428)       0:00:39.631 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  02:49:40 +0000 (0:00:00.401)       0:00:40.032 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  02:49:40 +0000 (0:00:00.355)       0:00:40.388 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  02:49:41 +0000 (0:00:01.045)       0:00:41.433 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  02:49:42 +0000 (0:00:00.821)       0:00:42.255 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:49:44 +0000 (0:00:02.014)       0:00:44.269 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:49:46 +0000 (0:00:01.312)       0:00:45.582 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:49:47 +0000 (0:00:01.295)       0:00:46.877 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:49:48 +0000 (0:00:01.274)       0:00:48.152 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:49:49 +0000 (0:00:00.747)       0:00:48.900 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:49:50 +0000 (0:00:01.056)       0:00:49.957 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:49:51 +0000 (0:00:00.725)       0:00:50.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:49:52 +0000 (0:00:01.363)       0:00:52.046 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:49:53 +0000 (0:00:00.664)       0:00:52.710 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:49:54 +0000 (0:00:00.893)       0:00:53.604 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:49:55 +0000 (0:00:01.075)       0:00:54.679 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:49:56 +0000 (0:00:01.406)       0:00:56.086 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:49:57 +0000 (0:00:00.657)       0:00:56.744 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:49:58 +0000 (0:00:00.990)       0:00:57.735 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:49:58 +0000 (0:00:00.682)       0:00:58.417 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:49:59 +0000 (0:00:00.876)       0:00:59.293 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:50:00 +0000 (0:00:00.609)       0:00:59.903 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:50:01 +0000 (0:00:00.898)       0:01:00.801 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:50:01 +0000 (0:00:00.603)       0:01:01.405 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:50:02 +0000 (0:00:00.919)       0:01:02.324 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:50:03 +0000 (0:00:00.998)       0:01:03.322 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  02:50:05 +0000 (0:00:01.382)       0:01:04.705 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  02:50:06 +0000 (0:00:01.512)       0:01:06.217 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  02:50:08 +0000 (0:00:02.252)       0:01:08.470 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  02:50:10 +0000 (0:00:01.342)       0:01:09.812 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  02:50:10 +0000 (0:00:00.682)       0:01:10.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  02:50:11 +0000 (0:00:00.966)       0:01:11.462 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  02:50:13 +0000 (0:00:01.348)       0:01:12.810 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  02:50:14 +0000 (0:00:01.234)       0:01:14.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  02:50:15 +0000 (0:00:00.679)       0:01:14.725 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  02:50:16 +0000 (0:00:00.941)       0:01:15.667 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  02:50:16 +0000 (0:00:00.626)       0:01:16.293 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  02:50:18 +0000 (0:00:01.326)       0:01:17.620 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  02:51:22 +0000 (0:01:04.165)       0:02:21.785 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  02:51:23 +0000 (0:00:00.742)       0:02:22.528 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  02:51:23 +0000 (0:00:00.947)       0:02:23.475 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  02:51:24 +0000 (0:00:00.167)       0:02:23.643 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  02:51:24 +0000 (0:00:00.674)       0:02:24.318 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  02:51:25 +0000 (0:00:00.881)       0:02:25.199 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  02:51:26 +0000 (0:00:00.695)       0:02:25.894 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  02:51:29 +0000 (0:00:03.470)       0:02:29.365 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  02:51:30 +0000 (0:00:00.685)       0:02:30.050 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  02:51:31 +0000 (0:00:01.322)       0:02:31.372 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  02:51:32 +0000 (0:00:00.904)       0:02:32.276 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  02:51:33 +0000 (0:00:00.626)       0:02:32.903 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  02:51:34 +0000 (0:00:00.898)       0:02:33.801 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  02:51:34 +0000 (0:00:00.702)       0:02:34.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  02:51:35 +0000 (0:00:00.898)       0:02:35.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  02:51:36 +0000 (0:00:00.664)       0:02:36.066 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  02:51:37 +0000 (0:00:00.963)       0:02:37.030 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  02:51:38 +0000 (0:00:00.915)       0:02:37.945 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  02:51:39 +0000 (0:00:01.391)       0:02:39.337 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  02:51:41 +0000 (0:00:01.321)       0:02:40.658 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  02:51:47 +0000 (0:00:06.356)       0:02:47.015 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  02:51:48 +0000 (0:00:01.133)       0:02:48.148 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  02:51:49 +0000 (0:00:00.745)       0:02:48.894 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  02:51:50 +0000 (0:00:01.240)       0:02:50.134 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  02:51:51 +0000 (0:00:00.920)       0:02:51.054 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  02:51:52 +0000 (0:00:00.665)       0:02:51.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  02:51:53 +0000 (0:00:00.889)       0:02:52.609 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  02:51:53 +0000 (0:00:00.631)       0:02:53.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  02:51:54 +0000 (0:00:00.939)       0:02:54.180 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  02:51:55 +0000 (0:00:00.688)       0:02:54.869 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  02:51:56 +0000 (0:00:00.934)       0:02:55.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  02:51:56 +0000 (0:00:00.685)       0:02:56.489 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  02:51:59 +0000 (0:00:02.982)       0:02:59.471 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  02:52:00 +0000 (0:00:00.688)       0:03:00.159 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  02:52:01 +0000 (0:00:00.963)       0:03:01.122 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  02:52:02 +0000 (0:00:00.895)       0:03:02.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  02:52:03 +0000 (0:00:00.702)       0:03:02.720 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  02:52:04 +0000 (0:00:01.098)       0:03:03.818 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  02:52:04 +0000 (0:00:00.647)       0:03:04.466 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  02:52:05 +0000 (0:00:00.871)       0:03:05.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  02:52:06 +0000 (0:00:00.713)       0:03:06.051 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  02:52:07 +0000 (0:00:00.860)       0:03:06.912 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  02:52:08 +0000 (0:00:00.682)       0:03:07.594 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  02:52:08 +0000 (0:00:00.894)       0:03:08.489 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  02:52:09 +0000 (0:00:00.460)       0:03:08.949 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  02:52:09 +0000 (0:00:00.447)       0:03:09.396 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  02:52:10 +0000 (0:00:00.466)       0:03:09.863 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  02:52:11 +0000 (0:00:00.670)       0:03:10.534 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  02:52:12 +0000 (0:00:01.812)       0:03:12.346 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  02:52:15 +0000 (0:00:02.709)       0:03:15.056 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  02:52:16 +0000 (0:00:01.012)       0:03:16.069 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:52:17 +0000 (0:00:01.151)       0:03:17.220 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Thursday 26 March 2026  02:52:18 +0000 (0:00:00.329)       0:03:17.550 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Thursday 26 March 2026  02:52:19 +0000 (0:00:01.531)       0:03:19.081 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:52:20 +0000 (0:00:00.693)       0:03:19.775 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  02:52:20 +0000 (0:00:00.363)       0:03:20.139 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Thursday 26 March 2026  02:52:21 +0000 (0:00:01.200)       0:03:21.340 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:52:22 +0000 (0:00:00.429)       0:03:21.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Thursday 26 March 2026  02:52:22 +0000 (0:00:00.336)       0:03:22.105 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Thursday 26 March 2026  02:52:22 +0000 (0:00:00.246)       0:03:22.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Thursday 26 March 2026  02:52:23 +0000 (0:00:00.546)       0:03:22.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Thursday 26 March 2026  02:52:23 +0000 (0:00:00.261)       0:03:23.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Thursday 26 March 2026  02:52:23 +0000 (0:00:00.255)       0:03:23.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Thursday 26 March 2026  02:52:24 +0000 (0:00:00.150)       0:03:23.567 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Thursday 26 March 2026  02:52:24 +0000 (0:00:00.259)       0:03:23.827 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Thursday 26 March 2026  02:52:24 +0000 (0:00:00.238)       0:03:24.065 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:52:25 +0000 (0:00:00.455)       0:03:24.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Thursday 26 March 2026  02:52:25 +0000 (0:00:00.331)       0:03:24.852 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Thursday 26 March 2026  02:52:25 +0000 (0:00:00.265)       0:03:25.117 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  02:52:26 +0000 (0:00:00.782)       0:03:25.900 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:52:27 +0000 (0:00:00.893)       0:03:26.794 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Thursday 26 March 2026  02:52:27 +0000 (0:00:00.562)       0:03:27.356 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Thursday 26 March 2026  02:52:29 +0000 (0:00:01.284)       0:03:28.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:52:29 +0000 (0:00:00.679)       0:03:29.321 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  02:52:30 +0000 (0:00:00.331)       0:03:29.652 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:52:31 +0000 (0:00:01.196)       0:03:30.849 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Thursday 26 March 2026  02:52:31 +0000 (0:00:00.369)       0:03:31.219 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Thursday 26 March 2026  02:52:32 +0000 (0:00:01.245)       0:03:32.465 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:52:34 +0000 (0:00:01.165)       0:03:33.630 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  02:52:34 +0000 (0:00:00.386)       0:03:34.017 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  02:52:35 +0000 (0:00:00.727)       0:03:34.744 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:52:36 +0000 (0:00:01.149)       0:03:35.894 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Thursday 26 March 2026  02:52:36 +0000 (0:00:00.381)       0:03:36.275 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Thursday 26 March 2026  02:52:38 +0000 (0:00:01.464)       0:03:37.740 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:52:38 +0000 (0:00:00.659)       0:03:38.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mon] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:52:39 +0000 (0:00:00.630)       0:03:39.030 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:52:40 +0000 (0:00:00.811)       0:03:39.841 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:52:40 +0000 (0:00:00.567)       0:03:40.409 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:52:41 +0000 (0:00:00.741)       0:03:41.151 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:52:42 +0000 (0:00:00.587)       0:03:41.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:52:42 +0000 (0:00:00.375)       0:03:42.114 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:52:42 +0000 (0:00:00.317)       0:03:42.431 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:52:43 +0000 (0:00:00.729)       0:03:43.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:52:44 +0000 (0:00:00.626)       0:03:43.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:52:44 +0000 (0:00:00.399)       0:03:44.187 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:52:45 +0000 (0:00:00.853)       0:03:45.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:52:46 +0000 (0:00:00.753)       0:03:45.793 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:52:46 +0000 (0:00:00.609)       0:03:46.403 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:52:47 +0000 (0:00:00.373)       0:03:46.776 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:52:47 +0000 (0:00:00.324)       0:03:47.101 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:52:48 +0000 (0:00:00.576)       0:03:47.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:52:48 +0000 (0:00:00.356)       0:03:48.033 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:52:48 +0000 (0:00:00.355)       0:03:48.389 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:52:49 +0000 (0:00:00.348)       0:03:48.737 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:52:49 +0000 (0:00:00.666)       0:03:49.404 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:52:50 +0000 (0:00:00.379)       0:03:49.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  02:52:50 +0000 (0:00:00.593)       0:03:50.377 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Thursday 26 March 2026  02:52:51 +0000 (0:00:00.624)       0:03:51.002 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Thursday 26 March 2026  02:52:52 +0000 (0:00:00.610)       0:03:51.613 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Thursday 26 March 2026  02:52:52 +0000 (0:00:00.156)       0:03:51.769 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Thursday 26 March 2026  02:52:53 +0000 (0:00:01.067)       0:03:52.836 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Thursday 26 March 2026  02:52:53 +0000 (0:00:00.613)       0:03:53.450 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Thursday 26 March 2026  02:52:54 +0000 (0:00:00.378)       0:03:53.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Thursday 26 March 2026  02:52:55 +0000 (0:00:01.161)       0:03:54.990 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Thursday 26 March 2026  02:52:56 +0000 (0:00:00.830)       0:03:55.820 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Thursday 26 March 2026  02:52:57 +0000 (0:00:00.950)       0:03:56.771 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Thursday 26 March 2026  02:52:57 +0000 (0:00:00.730)       0:03:57.501 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Thursday 26 March 2026  02:52:59 +0000 (0:00:01.250)       0:03:58.752 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Thursday 26 March 2026  02:52:59 +0000 (0:00:00.730)       0:03:59.483 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Thursday 26 March 2026  02:53:03 +0000 (0:00:03.125)       0:04:02.608 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Thursday 26 March 2026  02:53:04 +0000 (0:00:01.266)       0:04:03.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Thursday 26 March 2026  02:53:05 +0000 (0:00:00.657)       0:04:04.532 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Thursday 26 March 2026  02:53:05 +0000 (0:00:00.355)       0:04:04.887 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Thursday 26 March 2026  02:53:06 +0000 (0:00:01.533)       0:04:06.420 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Thursday 26 March 2026  02:53:08 +0000 (0:00:01.632)       0:04:08.052 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Thursday 26 March 2026  02:53:08 +0000 (0:00:00.334)       0:04:08.387 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  02:53:09 +0000 (0:00:00.567)       0:04:08.954 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Thursday 26 March 2026  02:53:10 +0000 (0:00:00.631)       0:04:09.585 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  02:53:10 +0000 (0:00:00.350)       0:04:09.936 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Thursday 26 March 2026  02:53:11 +0000 (0:00:00.590)       0:04:10.527 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Thursday 26 March 2026  02:53:13 +0000 (0:00:02.129)       0:04:12.656 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Thursday 26 March 2026  02:53:14 +0000 (0:00:01.196)       0:04:13.852 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Thursday 26 March 2026  02:53:16 +0000 (0:00:01.859)       0:04:15.712 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Thursday 26 March 2026  02:53:18 +0000 (0:00:01.955)       0:04:17.667 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Thursday 26 March 2026  02:53:19 +0000 (0:00:00.897)       0:04:18.565 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Thursday 26 March 2026  02:53:40 +0000 (0:00:21.851)       0:04:40.416 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Thursday 26 March 2026  02:53:49 +0000 (0:00:08.986)       0:04:49.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Thursday 26 March 2026  02:53:50 +0000 (0:00:00.363)       0:04:49.766 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__e255bad6a7d49c7c14086d2eafbc7336e14b386d'}]) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  02:54:05 +0000 (0:00:14.915)       0:05:04.681 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  02:54:05 +0000 (0:00:00.452)       0:05:05.134 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:54:06 +0000 (0:00:00.870)       0:05:06.004 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Thursday 26 March 2026  02:54:06 +0000 (0:00:00.357)       0:05:06.362 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Thursday 26 March 2026  02:54:07 +0000 (0:00:00.385)       0:05:06.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:54:08 +0000 (0:00:00.959)       0:05:07.707 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mgr] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:54:09 +0000 (0:00:00.899)       0:05:08.607 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:54:09 +0000 (0:00:00.580)       0:05:09.188 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:54:10 +0000 (0:00:00.840)       0:05:10.028 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:54:11 +0000 (0:00:00.768)       0:05:10.796 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:54:11 +0000 (0:00:00.352)       0:05:11.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:54:12 +0000 (0:00:00.577)       0:05:11.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:54:12 +0000 (0:00:00.354)       0:05:12.081 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:54:13 +0000 (0:00:00.788)       0:05:12.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:54:13 +0000 (0:00:00.335)       0:05:13.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:54:14 +0000 (0:00:00.644)       0:05:13.850 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:54:15 +0000 (0:00:00.770)       0:05:14.621 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:54:15 +0000 (0:00:00.733)       0:05:15.354 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:54:16 +0000 (0:00:00.328)       0:05:15.683 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:54:16 +0000 (0:00:00.665)       0:05:16.348 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:54:17 +0000 (0:00:00.390)       0:05:16.739 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:54:17 +0000 (0:00:00.336)       0:05:17.075 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:54:18 +0000 (0:00:00.613)       0:05:17.689 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:54:18 +0000 (0:00:00.324)       0:05:18.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:54:18 +0000 (0:00:00.341)       0:05:18.355 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:54:19 +0000 (0:00:00.379)       0:05:18.735 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:54:19 +0000 (0:00:00.632)       0:05:19.368 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  02:54:20 +0000 (0:00:00.601)       0:05:19.969 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  02:54:21 +0000 (0:00:00.911)       0:05:20.881 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Thursday 26 March 2026  02:54:22 +0000 (0:00:00.805)       0:05:21.686 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Thursday 26 March 2026  02:54:22 +0000 (0:00:00.742)       0:05:22.429 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Thursday 26 March 2026  02:54:23 +0000 (0:00:00.355)       0:05:22.784 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Thursday 26 March 2026  02:54:33 +0000 (0:00:10.631)       0:05:33.415 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  02:54:34 +0000 (0:00:00.641)       0:05:34.056 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  02:54:36 +0000 (0:00:02.303)       0:05:36.360 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Thursday 26 March 2026  02:54:38 +0000 (0:00:01.397)       0:05:37.758 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Thursday 26 March 2026  02:54:39 +0000 (0:00:00.819)       0:05:38.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  02:54:39 +0000 (0:00:00.606)       0:05:39.184 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  02:54:40 +0000 (0:00:00.359)       0:05:39.544 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  02:54:40 +0000 (0:00:00.604)       0:05:40.149 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Thursday 26 March 2026  02:54:41 +0000 (0:00:00.608)       0:05:40.757 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  02:54:41 +0000 (0:00:00.367)       0:05:41.125 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  02:54:42 +0000 (0:00:00.548)       0:05:41.674 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Thursday 26 March 2026  02:54:44 +0000 (0:00:01.999)       0:05:43.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Thursday 26 March 2026  02:54:45 +0000 (0:00:01.225)       0:05:44.898 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Thursday 26 March 2026  02:54:47 +0000 (0:00:01.786)       0:05:46.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  02:54:49 +0000 (0:00:02.736)       0:05:49.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Thursday 26 March 2026  02:54:50 +0000 (0:00:00.728)       0:05:50.149 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Thursday 26 March 2026  02:55:14 +0000 (0:00:24.280)       0:06:14.429 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Thursday 26 March 2026  02:55:16 +0000 (0:00:01.301)       0:06:15.731 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Thursday 26 March 2026  02:55:16 +0000 (0:00:00.330)       0:06:16.061 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Thursday 26 March 2026  02:55:16 +0000 (0:00:00.174)       0:06:16.236 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Thursday 26 March 2026  02:55:23 +0000 (0:00:06.365)       0:06:22.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  02:55:28 +0000 (0:00:05.322)       0:06:27.925 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  02:55:29 +0000 (0:00:00.706)       0:06:28.631 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:55:30 +0000 (0:00:00.903)       0:06:29.534 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Thursday 26 March 2026  02:55:30 +0000 (0:00:00.370)       0:06:29.905 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Thursday 26 March 2026  02:55:31 +0000 (0:00:01.162)       0:06:31.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:55:32 +0000 (0:00:00.920)       0:06:31.989 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-osd] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:55:33 +0000 (0:00:00.895)       0:06:32.884 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:55:33 +0000 (0:00:00.594)       0:06:33.479 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:55:34 +0000 (0:00:00.808)       0:06:34.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:55:35 +0000 (0:00:00.389)       0:06:34.678 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:55:35 +0000 (0:00:00.778)       0:06:35.457 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:55:36 +0000 (0:00:00.694)       0:06:36.152 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:55:37 +0000 (0:00:01.077)       0:06:37.229 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:55:38 +0000 (0:00:00.346)       0:06:37.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:55:38 +0000 (0:00:00.395)       0:06:37.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:55:38 +0000 (0:00:00.328)       0:06:38.300 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:55:39 +0000 (0:00:01.064)       0:06:39.364 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:55:40 +0000 (0:00:00.726)       0:06:40.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:55:40 +0000 (0:00:00.328)       0:06:40.419 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:55:41 +0000 (0:00:00.360)       0:06:40.780 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:55:41 +0000 (0:00:00.663)       0:06:41.444 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:55:42 +0000 (0:00:00.355)       0:06:41.800 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:55:42 +0000 (0:00:00.384)       0:06:42.184 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:55:43 +0000 (0:00:00.333)       0:06:42.518 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:55:43 +0000 (0:00:00.651)       0:06:43.169 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:55:43 +0000 (0:00:00.343)       0:06:43.513 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:55:44 +0000 (0:00:00.361)       0:06:43.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Thursday 26 March 2026  02:55:45 +0000 (0:00:00.879)       0:06:44.754 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  02:55:45 +0000 (0:00:00.347)       0:06:45.101 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Thursday 26 March 2026  02:55:46 +0000 (0:00:00.914)       0:06:46.015 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Thursday 26 March 2026  02:55:47 +0000 (0:00:00.837)       0:06:46.853 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Thursday 26 March 2026  02:55:47 +0000 (0:00:00.338)       0:06:47.191 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Thursday 26 March 2026  02:55:48 +0000 (0:00:00.355)       0:06:47.547 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Thursday 26 March 2026  02:55:48 +0000 (0:00:00.654)       0:06:48.202 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Thursday 26 March 2026  02:55:49 +0000 (0:00:00.653)       0:06:48.855 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Thursday 26 March 2026  02:55:51 +0000 (0:00:01.974)       0:06:50.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Thursday 26 March 2026  02:55:51 +0000 (0:00:00.334)       0:06:51.165 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Thursday 26 March 2026  02:55:52 +0000 (0:00:00.852)       0:06:52.017 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  02:55:53 +0000 (0:00:01.019)       0:06:53.037 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  02:55:55 +0000 (0:00:02.033)       0:06:55.070 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Thursday 26 March 2026  02:55:56 +0000 (0:00:01.153)       0:06:56.224 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Thursday 26 March 2026  02:55:58 +0000 (0:00:02.086)       0:06:58.311 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Use ceph-volume to create osds] *******************************",
                                "Thursday 26 March 2026  02:55:59 +0000 (0:00:00.966)       0:06:59.277 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Thursday 26 March 2026  02:56:40 +0000 (0:00:41.005)       0:07:40.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Thursday 26 March 2026  02:56:41 +0000 (0:00:00.358)       0:07:40.642 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Thursday 26 March 2026  02:56:41 +0000 (0:00:00.842)       0:07:41.484 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Thursday 26 March 2026  02:56:42 +0000 (0:00:00.722)       0:07:42.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  02:56:45 +0000 (0:00:02.698)       0:07:44.905 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  02:56:46 +0000 (0:00:00.912)       0:07:45.817 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Thursday 26 March 2026  02:56:47 +0000 (0:00:01.197)       0:07:47.014 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Thursday 26 March 2026  02:56:48 +0000 (0:00:01.249)       0:07:48.264 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  02:56:50 +0000 (0:00:02.004)       0:07:50.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Thursday 26 March 2026  02:56:51 +0000 (0:00:00.340)       0:07:50.609 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Thursday 26 March 2026  02:56:51 +0000 (0:00:00.352)       0:07:50.962 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Thursday 26 March 2026  02:56:52 +0000 (0:00:01.029)       0:07:51.991 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Thursday 26 March 2026  02:56:54 +0000 (0:00:02.493)       0:07:54.484 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Thursday 26 March 2026  02:56:58 +0000 (0:00:03.433)       0:07:57.918 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Thursday 26 March 2026  02:57:01 +0000 (0:00:03.095)       0:08:01.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Thursday 26 March 2026  02:57:14 +0000 (0:00:12.703)       0:08:13.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  02:57:15 +0000 (0:00:01.343)       0:08:15.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  02:57:16 +0000 (0:00:00.684)       0:08:15.745 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Thursday 26 March 2026  02:57:16 +0000 (0:00:00.569)       0:08:16.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:57:17 +0000 (0:00:00.415)       0:08:16.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Thursday 26 March 2026  02:57:17 +0000 (0:00:00.374)       0:08:17.105 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Thursday 26 March 2026  02:57:17 +0000 (0:00:00.252)       0:08:17.358 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Thursday 26 March 2026  02:57:18 +0000 (0:00:00.733)       0:08:18.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Thursday 26 March 2026  02:57:18 +0000 (0:00:00.263)       0:08:18.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Thursday 26 March 2026  02:57:19 +0000 (0:00:00.258)       0:08:18.614 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Thursday 26 March 2026  02:57:19 +0000 (0:00:00.163)       0:08:18.777 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Thursday 26 March 2026  02:57:19 +0000 (0:00:00.246)       0:08:19.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Thursday 26 March 2026  02:57:19 +0000 (0:00:00.263)       0:08:19.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:57:20 +0000 (0:00:00.432)       0:08:19.720 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Thursday 26 March 2026  02:57:20 +0000 (0:00:00.640)       0:08:20.361 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Thursday 26 March 2026  02:57:21 +0000 (0:00:00.259)       0:08:20.621 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-crash] ***************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:57:21 +0000 (0:00:00.758)       0:08:21.380 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:57:23 +0000 (0:00:01.348)       0:08:22.728 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:57:24 +0000 (0:00:01.484)       0:08:24.213 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:57:26 +0000 (0:00:01.323)       0:08:25.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:57:26 +0000 (0:00:00.783)       0:08:26.320 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:57:27 +0000 (0:00:00.990)       0:08:27.311 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:57:28 +0000 (0:00:00.724)       0:08:28.035 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:57:29 +0000 (0:00:01.419)       0:08:29.455 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:57:30 +0000 (0:00:00.658)       0:08:30.113 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:57:31 +0000 (0:00:00.965)       0:08:31.079 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:57:32 +0000 (0:00:01.045)       0:08:32.124 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:57:34 +0000 (0:00:01.424)       0:08:33.549 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:57:34 +0000 (0:00:00.668)       0:08:34.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:57:35 +0000 (0:00:00.985)       0:08:35.203 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:57:36 +0000 (0:00:00.647)       0:08:35.851 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:57:37 +0000 (0:00:00.949)       0:08:36.801 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:57:37 +0000 (0:00:00.606)       0:08:37.407 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:57:38 +0000 (0:00:00.962)       0:08:38.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:57:39 +0000 (0:00:00.674)       0:08:39.044 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:57:40 +0000 (0:00:00.913)       0:08:39.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:57:41 +0000 (0:00:00.682)       0:08:40.639 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Thursday 26 March 2026  02:57:42 +0000 (0:00:01.459)       0:08:42.099 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Thursday 26 March 2026  02:57:46 +0000 (0:00:04.199)       0:08:46.299 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Thursday 26 March 2026  02:57:48 +0000 (0:00:01.847)       0:08:48.147 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Thursday 26 March 2026  02:57:50 +0000 (0:00:01.475)       0:08:49.623 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Thursday 26 March 2026  02:57:51 +0000 (0:00:01.315)       0:08:50.938 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Thursday 26 March 2026  02:57:52 +0000 (0:00:01.345)       0:08:52.284 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Thursday 26 March 2026  02:57:54 +0000 (0:00:01.481)       0:08:53.765 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************",
                                "Thursday 26 March 2026  02:57:57 +0000 (0:00:03.720)       0:08:57.486 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******",
                                "Thursday 26 March 2026  02:57:59 +0000 (0:00:01.429)       0:08:58.915 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************",
                                "Thursday 26 March 2026  02:58:00 +0000 (0:00:00.769)       0:08:59.685 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******",
                                "Thursday 26 March 2026  02:58:02 +0000 (0:00:02.440)       0:09:02.125 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mds] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:58:03 +0000 (0:00:01.349)       0:09:03.475 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:58:04 +0000 (0:00:00.584)       0:09:04.059 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:58:05 +0000 (0:00:00.894)       0:09:04.954 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:58:05 +0000 (0:00:00.345)       0:09:05.299 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:58:06 +0000 (0:00:00.765)       0:09:06.065 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:58:07 +0000 (0:00:00.668)       0:09:06.733 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:58:08 +0000 (0:00:01.029)       0:09:07.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:58:08 +0000 (0:00:00.348)       0:09:08.111 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:58:08 +0000 (0:00:00.337)       0:09:08.449 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:58:09 +0000 (0:00:00.331)       0:09:08.780 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:58:10 +0000 (0:00:01.088)       0:09:09.869 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:58:11 +0000 (0:00:00.723)       0:09:10.592 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:58:11 +0000 (0:00:00.355)       0:09:10.947 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:58:11 +0000 (0:00:00.362)       0:09:11.309 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:58:12 +0000 (0:00:00.651)       0:09:11.961 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:58:12 +0000 (0:00:00.372)       0:09:12.333 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:58:13 +0000 (0:00:00.373)       0:09:12.707 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:58:13 +0000 (0:00:00.353)       0:09:13.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:58:14 +0000 (0:00:00.616)       0:09:13.678 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:58:14 +0000 (0:00:00.338)       0:09:14.016 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:58:14 +0000 (0:00:00.352)       0:09:14.369 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Thursday 26 March 2026  02:58:15 +0000 (0:00:00.853)       0:09:15.223 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule details] *********************",
                                "Thursday 26 March 2026  02:58:16 +0000 (0:00:00.477)       0:09:15.701 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule name] ************************",
                                "Thursday 26 March 2026  02:58:18 +0000 (0:00:02.073)       0:09:17.775 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create filesystem pools] **************************************",
                                "Thursday 26 March 2026  02:58:18 +0000 (0:00:00.244)       0:09:18.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create ceph filesystem] ***************************************",
                                "Thursday 26 March 2026  02:58:26 +0000 (0:00:08.350)       0:09:26.371 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  02:58:31 +0000 (0:00:04.352)       0:09:30.723 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Thursday 26 March 2026  02:58:31 +0000 (0:00:00.718)       0:09:31.442 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  02:58:32 +0000 (0:00:01.072)       0:09:32.514 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  02:58:35 +0000 (0:00:02.118)       0:09:34.633 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Thursday 26 March 2026  02:58:36 +0000 (0:00:01.512)       0:09:36.145 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Thursday 26 March 2026  02:58:39 +0000 (0:00:02.526)       0:09:38.672 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Thursday 26 March 2026  02:58:39 +0000 (0:00:00.370)       0:09:39.042 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  02:58:40 +0000 (0:00:00.954)       0:09:39.997 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  02:58:41 +0000 (0:00:00.669)       0:09:40.667 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Thursday 26 March 2026  02:58:42 +0000 (0:00:01.277)       0:09:41.944 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Thursday 26 March 2026  02:58:43 +0000 (0:00:01.458)       0:09:43.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Thursday 26 March 2026  02:58:45 +0000 (0:00:01.674)       0:09:45.077 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Thursday 26 March 2026  02:58:47 +0000 (0:00:01.827)       0:09:46.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  02:58:48 +0000 (0:00:01.559)       0:09:48.464 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  02:58:49 +0000 (0:00:00.669)       0:09:49.133 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Thursday 26 March 2026  02:58:50 +0000 (0:00:00.864)       0:09:49.998 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Thursday 26 March 2026  02:58:50 +0000 (0:00:00.370)       0:09:50.368 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Thursday 26 March 2026  02:58:51 +0000 (0:00:01.114)       0:09:51.482 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Thursday 26 March 2026  02:58:52 +0000 (0:00:00.959)       0:09:52.442 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  02:58:53 +0000 (0:00:00.926)       0:09:53.368 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  02:58:54 +0000 (0:00:00.558)       0:09:53.926 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  02:58:55 +0000 (0:00:00.896)       0:09:54.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  02:58:55 +0000 (0:00:00.342)       0:09:55.166 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  02:58:56 +0000 (0:00:00.682)       0:09:55.848 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  02:58:57 +0000 (0:00:01.071)       0:09:56.920 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  02:58:58 +0000 (0:00:00.699)       0:09:57.620 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  02:58:58 +0000 (0:00:00.347)       0:09:57.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  02:58:58 +0000 (0:00:00.337)       0:09:58.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  02:58:59 +0000 (0:00:00.669)       0:09:58.975 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  02:59:00 +0000 (0:00:00.714)       0:09:59.689 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  02:59:00 +0000 (0:00:00.704)       0:10:00.394 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  02:59:01 +0000 (0:00:00.354)       0:10:00.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  02:59:01 +0000 (0:00:00.652)       0:10:01.400 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  02:59:02 +0000 (0:00:00.372)       0:10:01.773 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  02:59:02 +0000 (0:00:00.370)       0:10:02.143 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  02:59:03 +0000 (0:00:00.378)       0:10:02.522 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  02:59:03 +0000 (0:00:00.658)       0:10:03.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  02:59:03 +0000 (0:00:00.328)       0:10:03.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  02:59:04 +0000 (0:00:00.358)       0:10:03.868 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  02:59:04 +0000 (0:00:00.338)       0:10:04.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  02:59:05 +0000 (0:00:00.940)       0:10:05.147 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  02:59:06 +0000 (0:00:00.846)       0:10:05.993 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  02:59:08 +0000 (0:00:02.156)       0:10:08.150 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Thursday 26 March 2026  02:59:09 +0000 (0:00:01.227)       0:10:09.378 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Thursday 26 March 2026  02:59:10 +0000 (0:00:00.373)       0:10:09.751 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Thursday 26 March 2026  02:59:11 +0000 (0:00:00.841)       0:10:10.593 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Thursday 26 March 2026  02:59:11 +0000 (0:00:00.815)       0:10:11.408 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  02:59:16 +0000 (0:00:04.560)       0:10:15.969 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  02:59:18 +0000 (0:00:02.437)       0:10:18.407 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Thursday 26 March 2026  02:59:20 +0000 (0:00:01.617)       0:10:20.024 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Thursday 26 March 2026  02:59:20 +0000 (0:00:00.267)       0:10:20.292 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Thursday 26 March 2026  02:59:21 +0000 (0:00:00.660)       0:10:20.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Thursday 26 March 2026  02:59:22 +0000 (0:00:00.660)       0:10:21.614 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Thursday 26 March 2026  02:59:53 +0000 (0:00:31.288)       0:10:52.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Thursday 26 March 2026  02:59:53 +0000 (0:00:00.365)       0:10:53.268 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Thursday 26 March 2026  02:59:54 +0000 (0:00:00.610)       0:10:53.878 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Thursday 26 March 2026  02:59:55 +0000 (0:00:00.648)       0:10:54.526 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  02:59:55 +0000 (0:00:00.915)       0:10:55.442 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Thursday 26 March 2026  02:59:57 +0000 (0:00:01.331)       0:10:56.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Thursday 26 March 2026  02:59:58 +0000 (0:00:01.234)       0:10:58.008 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Thursday 26 March 2026  03:00:00 +0000 (0:00:01.817)       0:10:59.825 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  03:00:03 +0000 (0:00:02.837)       0:11:02.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  03:00:03 +0000 (0:00:00.373)       0:11:03.036 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Thursday 26 March 2026  03:00:04 +0000 (0:00:00.907)       0:11:03.943 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Thursday 26 March 2026  03:00:04 +0000 (0:00:00.422)       0:11:04.365 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Thursday 26 March 2026  03:00:05 +0000 (0:00:00.521)       0:11:04.887 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Thursday 26 March 2026  03:00:06 +0000 (0:00:01.352)       0:11:06.239 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:00:06 +0000 (0:00:00.276)       0:11:06.515 ******** ",
                                "=============================================================================== ",
                                "ceph-container-common : Pulling Ceph container image ------------------- 64.17s",
                                "ceph-osd : Use ceph-volume to create osds ------------------------------ 41.01s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 31.29s",
                                "ceph-mgr : Wait for all mgr to be up ----------------------------------- 24.28s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.85s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 14.92s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 12.70s",
                                "ceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.63s",
                                "ceph-mon : Fetch ceph initial keys -------------------------------------- 8.99s",
                                "ceph-mds : Create filesystem pools -------------------------------------- 8.35s",
                                "ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.37s",
                                "ceph-config : Create ceph initial directories --------------------------- 6.36s",
                                "ceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.32s",
                                "ceph-rgw : Create rgw keyrings ------------------------------------------ 4.56s",
                                "ceph-mds : Create ceph filesystem --------------------------------------- 4.35s",
                                "ceph-crash : Create client.crash keyring -------------------------------- 4.20s",
                                "ceph-crash : Start the ceph-crash service ------------------------------- 3.72s",
                                "ceph-container-common : Get ceph version -------------------------------- 3.47s",
                                "ceph-osd : Systemd start osd -------------------------------------------- 3.43s",
                                "ceph-mon : Copy admin keyring over to mons ------------------------------ 3.13s",
                                "\u001b[32m2026-03-26 03:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62df255a-8406-4b80-a8bf-efaf3055d2ba (ceph-pools) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62df255a-8406-4b80-a8bf-efaf3055d2ba (ceph-pools) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Create ceph pools] *******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  03:00:15 +0000 (0:00:00.682)       0:00:00.682 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  03:00:16 +0000 (0:00:00.729)       0:00:01.411 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  03:00:16 +0000 (0:00:00.672)       0:00:02.083 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  03:00:17 +0000 (0:00:00.346)       0:00:02.430 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  03:00:18 +0000 (0:00:00.953)       0:00:03.384 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  03:00:18 +0000 (0:00:00.372)       0:00:03.756 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  03:00:18 +0000 (0:00:00.309)       0:00:04.066 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  03:00:19 +0000 (0:00:00.344)       0:00:04.410 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  03:00:19 +0000 (0:00:00.607)       0:00:05.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  03:00:19 +0000 (0:00:00.320)       0:00:05.339 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  03:00:20 +0000 (0:00:00.675)       0:00:06.014 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  03:00:21 +0000 (0:00:00.504)       0:00:06.518 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  03:00:23 +0000 (0:00:02.274)       0:00:08.793 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  03:00:24 +0000 (0:00:00.695)       0:00:09.488 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  03:00:25 +0000 (0:00:01.166)       0:00:10.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  03:00:25 +0000 (0:00:00.178)       0:00:10.833 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 03:00:22.131422', 'end': '2026-03-26 03:00:22.186955', 'delta': '0:00:00.055533', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 03:00:22.709658', 'end': '2026-03-26 03:00:22.757960', 'delta': '0:00:00.048302', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 03:00:23.269066', 'end': '2026-03-26 03:00:23.309660', 'delta': '0:00:00.040594', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  03:00:25 +0000 (0:00:00.199)       0:00:11.032 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  03:00:26 +0000 (0:00:00.455)       0:00:11.488 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  03:00:27 +0000 (0:00:01.685)       0:00:13.174 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  03:00:28 +0000 (0:00:00.362)       0:00:13.536 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  03:00:28 +0000 (0:00:00.741)       0:00:14.277 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  03:00:29 +0000 (0:00:00.345)       0:00:14.623 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  03:00:29 +0000 (0:00:00.162)       0:00:14.785 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  03:00:29 +0000 (0:00:00.263)       0:00:15.048 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  03:00:30 +0000 (0:00:00.378)       0:00:15.427 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  03:00:30 +0000 (0:00:00.601)       0:00:16.028 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  03:00:31 +0000 (0:00:00.353)       0:00:16.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  03:00:31 +0000 (0:00:00.370)       0:00:16.752 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  03:00:31 +0000 (0:00:00.344)       0:00:17.096 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  03:00:32 +0000 (0:00:00.537)       0:00:17.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  03:00:32 +0000 (0:00:00.358)       0:00:17.992 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  03:00:33 +0000 (0:00:00.693)       0:00:18.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-09-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-26-01-38-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  03:00:33 +0000 (0:00:00.638)       0:00:19.323 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  03:00:34 +0000 (0:00:01.004)       0:00:20.328 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  03:00:35 +0000 (0:00:00.323)       0:00:20.652 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  03:00:35 +0000 (0:00:00.673)       0:00:21.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  03:00:36 +0000 (0:00:00.312)       0:00:21.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  03:00:37 +0000 (0:00:00.703)       0:00:22.342 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  03:00:37 +0000 (0:00:00.372)       0:00:22.715 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  03:00:38 +0000 (0:00:01.134)       0:00:23.849 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  03:00:38 +0000 (0:00:00.419)       0:00:24.269 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  03:00:39 +0000 (0:00:00.802)       0:00:25.072 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  03:00:40 +0000 (0:00:00.331)       0:00:25.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  03:00:40 +0000 (0:00:00.329)       0:00:25.734 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  03:00:40 +0000 (0:00:00.595)       0:00:26.329 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  03:00:41 +0000 (0:00:00.440)       0:00:26.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  03:00:41 +0000 (0:00:00.407)       0:00:27.176 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  03:00:42 +0000 (0:00:00.375)       0:00:27.552 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  03:00:42 +0000 (0:00:00.426)       0:00:27.978 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  03:00:42 +0000 (0:00:00.347)       0:00:28.326 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  03:00:43 +0000 (0:00:00.839)       0:00:29.166 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  03:00:44 +0000 (0:00:00.872)       0:00:30.038 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Include tasks from the ceph-osd role] ************************************",
                                "Thursday 26 March 2026  03:00:46 +0000 (0:00:01.762)       0:00:31.801 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [create openstack pool(s)] ************************************************",
                                "Thursday 26 March 2026  03:00:47 +0000 (0:00:00.617)       0:00:32.418 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [generate keys] ***********************************************************",
                                "Thursday 26 March 2026  03:01:33 +0000 (0:00:46.350)       0:01:18.768 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [get keys from monitors] **************************************************",
                                "Thursday 26 March 2026  03:01:57 +0000 (0:00:23.918)       0:01:42.687 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [copy ceph key(s) if needed] **********************************************",
                                "Thursday 26 March 2026  03:02:08 +0000 (0:00:11.575)       0:01:54.262 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:02:26 +0000 (0:00:17.220)       0:02:11.483 ******** ",
                                "=============================================================================== ",
                                "create openstack pool(s) ----------------------------------------------- 46.35s",
                                "generate keys ---------------------------------------------------------- 23.92s",
                                "copy ceph key(s) if needed --------------------------------------------- 17.22s",
                                "get keys from monitors ------------------------------------------------- 11.58s",
                                "ceph-facts : Find a running mon container ------------------------------- 2.27s",
                                "ceph-facts : Set_fact ceph_admin_command -------------------------------- 1.76s",
                                "ceph-facts : Get current fsid if cluster is already running ------------- 1.69s",
                                "ceph-facts : Check if the ceph mon socket is in-use --------------------- 1.17s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 1.13s",
                                "ceph-facts : Check if the ceph conf exists ------------------------------ 1.00s",
                                "ceph-facts : Check if podman binary is present -------------------------- 0.95s",
                                "ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.87s",
                                "ceph-facts : Set_fact rgw_instances ------------------------------------- 0.84s",
                                "ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.80s",
                                "ceph-facts : Get current fsid ------------------------------------------- 0.74s",
                                "ceph-facts : Include facts.yml ------------------------------------------ 0.73s",
                                "ceph-facts : Read osd pool default crush rule --------------------------- 0.70s",
                                "ceph-facts : Check for a ceph mon socket -------------------------------- 0.70s",
                                "ceph-facts : Collect existed devices ------------------------------------ 0.69s",
                                "ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.68s",
                                "\u001b[32m2026-03-26 03:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba06c60f-34f0-4428-a23d-714319bfb192 (copy-ceph-keys) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ba06c60f-34f0-4428-a23d-714319bfb192 (copy-ceph-keys) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy ceph keys to the configuration repository] **************************",
                                "",
                                "TASK [Check if ceph keys exist] ************************************************",
                                "Thursday 26 March 2026  03:02:33 +0000 (0:00:00.164)       0:00:00.164 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Fetch all ceph keys] *****************************************************",
                                "Thursday 26 March 2026  03:02:38 +0000 (0:00:04.691)       0:00:04.855 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Create share directory] **************************************************",
                                "Thursday 26 March 2026  03:02:42 +0000 (0:00:04.424)       0:00:09.280 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the share directory] **********************************",
                                "Thursday 26 March 2026  03:02:44 +0000 (0:00:01.072)       0:00:10.353 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Check if target directories exist] ***************************************",
                                "Thursday 26 March 2026  03:02:58 +0000 (0:00:14.271)       0:00:24.624 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the configuration directory] **************************",
                                "Thursday 26 March 2026  03:03:02 +0000 (0:00:04.262)       0:00:28.887 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:03:09 +0000 (0:00:07.279)       0:00:36.166 ******** ",
                                "=============================================================================== ",
                                "Write ceph keys to the share directory --------------------------------- 14.27s",
                                "Write ceph keys to the configuration directory -------------------------- 7.28s",
                                "Check if ceph keys exist ------------------------------------------------ 4.69s",
                                "Fetch all ceph keys ----------------------------------------------------- 4.42s",
                                "Check if target directories exist --------------------------------------- 4.26s",
                                "Create share directory -------------------------------------------------- 1.07s",
                                "\u001b[32m2026-03-26 03:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3617dc68-3c70-4fe1-b924-b0629be23217 (cephclient) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3617dc68-3c70-4fe1-b924-b0629be23217 (cephclient) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Thursday 26 March 2026  03:03:27 +0000 (0:00:00.256)       0:00:00.256 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Thursday 26 March 2026  03:03:27 +0000 (0:00:00.265)       0:00:00.522 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Thursday 26 March 2026  03:03:28 +0000 (0:00:01.247)       0:00:01.769 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Thursday 26 March 2026  03:03:30 +0000 (0:00:01.495)       0:00:03.265 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Thursday 26 March 2026  03:03:31 +0000 (0:00:00.984)       0:00:04.249 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Thursday 26 March 2026  03:03:32 +0000 (0:00:00.973)       0:00:05.223 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Thursday 26 March 2026  03:04:15 +0000 (0:00:43.450)       0:00:48.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Thursday 26 March 2026  03:04:19 +0000 (0:00:04.299)       0:00:52.973 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Thursday 26 March 2026  03:04:20 +0000 (0:00:00.500)       0:00:53.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Thursday 26 March 2026  03:04:20 +0000 (0:00:00.158)       0:00:53.631 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******",
                                "Thursday 26 March 2026  03:04:21 +0000 (0:00:00.574)       0:00:54.205 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***",
                                "Thursday 26 March 2026  03:04:22 +0000 (0:00:01.680)       0:00:55.885 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******",
                                "Thursday 26 March 2026  03:04:23 +0000 (0:00:00.705)       0:00:56.591 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****",
                                "Thursday 26 March 2026  03:04:24 +0000 (0:00:00.622)       0:00:57.214 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:04:25 +0000 (0:00:01.594)       0:00:58.809 ******** ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Manage cephclient service ------------------ 43.45s",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 4.30s",
                                "osism.services.cephclient : Restart cephclient service ------------------ 1.68s",
                                "osism.services.cephclient : Copy bash completion scripts ---------------- 1.59s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.50s",
                                "osism.services.cephclient : Create required directories ----------------- 1.25s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.98s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.97s",
                                "osism.services.cephclient : Ensure that all containers are up ----------- 0.71s",
                                "osism.services.cephclient : Wait for an healthy service ----------------- 0.62s",
                                "osism.services.cephclient : Include rook task --------------------------- 0.57s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.50s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.27s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.16s",
                                "\u001b[32m2026-03-26 03:04:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccda482a-8362-4afe-bbda-d062fa374a95 (ceph-bootstrap-dashboard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:04:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ccda482a-8362-4afe-bbda-d062fa374a95 (ceph-bootstrap-dashboard) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Bootstraph ceph dashboard] ***********************************************",
                                "",
                                "TASK [Disable the ceph dashboard] **********************************************",
                                "Thursday 26 March 2026  03:04:33 +0000 (0:00:00.273)       0:00:00.273 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/ssl to false] ******************************************",
                                "Thursday 26 March 2026  03:04:35 +0000 (0:00:01.841)       0:00:02.115 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_port to 7000] ***********************************",
                                "Thursday 26 March 2026  03:04:36 +0000 (0:00:01.161)       0:00:03.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************",
                                "Thursday 26 March 2026  03:04:37 +0000 (0:00:01.144)       0:00:04.420 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_behaviour to error] ****************************",
                                "Thursday 26 March 2026  03:04:38 +0000 (0:00:01.271)       0:00:05.692 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_error_status_code to 404] **********************",
                                "Thursday 26 March 2026  03:04:40 +0000 (0:00:01.123)       0:00:06.816 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable the ceph dashboard] ***********************************************",
                                "Thursday 26 March 2026  03:04:41 +0000 (0:00:01.124)       0:00:07.941 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Write ceph_dashboard_password to temporary file] *************************",
                                "Thursday 26 March 2026  03:04:43 +0000 (0:00:02.107)       0:00:10.048 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create admin user] *******************************************************",
                                "Thursday 26 March 2026  03:04:44 +0000 (0:00:01.337)       0:00:11.386 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Remove temporary file for ceph_dashboard_password] ***********************",
                                "Thursday 26 March 2026  03:05:25 +0000 (0:00:40.591)       0:00:51.978 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Thursday 26 March 2026  03:05:25 +0000 (0:00:00.170)       0:00:52.149 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Thursday 26 March 2026  03:05:37 +0000 (0:00:12.040)       0:01:04.189 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Thursday 26 March 2026  03:05:38 +0000 (0:00:01.252)       0:01:05.441 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:05:50 +0000 (0:00:11.365)       0:01:16.807 ******** ",
                                "=============================================================================== ",
                                "Create admin user ------------------------------------------------------ 40.59s",
                                "Restart ceph manager service ------------------------------------------- 24.66s",
                                "Enable the ceph dashboard ----------------------------------------------- 2.11s",
                                "Disable the ceph dashboard ---------------------------------------------- 1.84s",
                                "Write ceph_dashboard_password to temporary file ------------------------- 1.34s",
                                "Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.27s",
                                "Set mgr/dashboard/ssl to false ------------------------------------------ 1.16s",
                                "Set mgr/dashboard/server_port to 7000 ----------------------------------- 1.14s",
                                "Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.12s",
                                "Set mgr/dashboard/standby_behaviour to error ---------------------------- 1.12s",
                                "Remove temporary file for ceph_dashboard_password ----------------------- 0.17s",
                                "+ sh -c /opt/configuration/scripts/deploy/300-openstack.sh",
                                "\u001b[32m2026-03-26 03:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5d3630be-9a4e-48be-9e20-a75775b77781 (keystone) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5d3630be-9a4e-48be-9e20-a75775b77781 (keystone) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:05:57 +0000 (0:00:00.282)       0:00:00.282 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:05:57 +0000 (0:00:00.343)       0:00:00.625 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Apply role keystone] *****************************************************",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:05:58 +0000 (0:00:00.442)       0:00:01.067 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Ensuring config directories exist] ****************************",
                                "Thursday 26 March 2026  03:05:58 +0000 (0:00:00.605)       0:00:01.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Check if policies shall be overwritten] ***********************",
                                "Thursday 26 March 2026  03:06:00 +0000 (0:00:01.621)       0:00:03.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Set keystone policy file] *************************************",
                                "Thursday 26 March 2026  03:06:00 +0000 (0:00:00.323)       0:00:03.618 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Check if Keystone domain-specific config is supplied] *********",
                                "Thursday 26 March 2026  03:06:01 +0000 (0:00:00.342)       0:00:03.961 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:06:01 +0000 (0:00:00.874)       0:00:04.835 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over extra CA certificates] *******",
                                "Thursday 26 March 2026  03:06:02 +0000 (0:00:00.623)       0:00:05.459 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:06:06 +0000 (0:00:03.477)       0:00:08.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****",
                                "Thursday 26 March 2026  03:06:06 +0000 (0:00:00.583)       0:00:09.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over config.json files for services] ******************",
                                "Thursday 26 March 2026  03:06:07 +0000 (0:00:00.888)       0:00:10.408 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone.conf] ***********************************",
                                "Thursday 26 March 2026  03:06:10 +0000 (0:00:03.322)       0:00:13.730 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying keystone-startup script for keystone] *****************",
                                "Thursday 26 March 2026  03:06:15 +0000 (0:00:04.719)       0:00:18.450 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Create Keystone domain-specific config directory] *************",
                                "Thursday 26 March 2026  03:06:16 +0000 (0:00:01.429)       0:00:19.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Get file list in custom domains folder] ***********************",
                                "Thursday 26 March 2026  03:06:17 +0000 (0:00:00.806)       0:00:20.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying Keystone Domain specific settings] ********************",
                                "Thursday 26 March 2026  03:06:18 +0000 (0:00:00.569)       0:00:21.256 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over existing policy file] ****************************",
                                "Thursday 26 March 2026  03:06:18 +0000 (0:00:00.323)       0:00:21.579 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:06:19 +0000 (0:00:00.588)       0:00:22.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over wsgi-keystone.conf] ******************************",
                                "Thursday 26 March 2026  03:06:19 +0000 (0:00:00.309)       0:00:22.477 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "",
                                "TASK [keystone : Checking whether keystone-paste.ini file exists] **************",
                                "Thursday 26 March 2026  03:06:21 +0000 (0:00:01.859)       0:00:24.336 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone-paste.ini] ******************************",
                                "Thursday 26 March 2026  03:06:22 +0000 (0:00:00.953)       0:00:25.290 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Generate the required cron jobs for the node] *****************",
                                "Thursday 26 March 2026  03:06:22 +0000 (0:00:00.601)       0:00:25.891 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***",
                                "Thursday 26 March 2026  03:06:24 +0000 (0:00:01.090)       0:00:26.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-fernet] ****************************",
                                "Thursday 26 March 2026  03:06:24 +0000 (0:00:00.565)       0:00:27.547 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-ssh] *******************************",
                                "Thursday 26 March 2026  03:06:33 +0000 (0:00:09.154)       0:00:36.701 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "",
                                "TASK [keystone : Check keystone containers] ************************************",
                                "Thursday 26 March 2026  03:06:36 +0000 (0:00:02.672)       0:00:39.373 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:06:38 +0000 (0:00:02.245)       0:00:41.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Creating keystone database] ***********************************",
                                "Thursday 26 March 2026  03:06:39 +0000 (0:00:00.588)       0:00:42.207 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating Keystone database user and setting permissions] ******",
                                "Thursday 26 March 2026  03:06:41 +0000 (0:00:02.362)       0:00:44.570 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Checking for any running keystone_fernet containers] **********",
                                "Thursday 26 March 2026  03:06:43 +0000 (0:00:02.231)       0:00:46.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Group nodes where keystone_fernet is running] *****************",
                                "Thursday 26 March 2026  03:06:44 +0000 (0:00:00.871)       0:00:47.673 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***",
                                "Thursday 26 March 2026  03:06:45 +0000 (0:00:00.388)       0:00:48.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone bootstrap container] *************************",
                                "Thursday 26 March 2026  03:06:45 +0000 (0:00:00.582)       0:00:48.644 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone fernet bootstrap container] ******************",
                                "Thursday 26 March 2026  03:07:00 +0000 (0:00:15.232)       0:01:03.876 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:07:12 +0000 (0:00:11.133)       0:01:15.010 ******** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:07:12 +0000 (0:00:00.085)       0:01:15.096 ******** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:07:12 +0000 (0:00:00.072)       0:01:15.169 ******** ",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-ssh container] ********************",
                                "Thursday 26 March 2026  03:07:12 +0000 (0:00:00.072)       0:01:15.241 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-fernet container] *****************",
                                "Thursday 26 March 2026  03:08:01 +0000 (0:00:49.106)       0:02:04.348 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone container] ************************",
                                "Thursday 26 March 2026  03:08:11 +0000 (0:00:10.443)       0:02:14.791 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:08:19 +0000 (0:00:07.404)       0:02:22.195 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Waiting for Keystone SSH port to be UP] ***********************",
                                "Thursday 26 March 2026  03:08:19 +0000 (0:00:00.644)       0:02:22.840 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Run key distribution] *****************************************",
                                "Thursday 26 March 2026  03:08:21 +0000 (0:00:01.178)       0:02:24.018 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating admin project, user, role, service, and endpoint] ****",
                                "Thursday 26 March 2026  03:08:22 +0000 (0:00:01.778)       0:02:25.797 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating services] **********************",
                                "Thursday 26 March 2026  03:08:34 +0000 (0:00:11.928)       0:02:37.725 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating endpoints] *********************",
                                "Thursday 26 March 2026  03:08:59 +0000 (0:00:24.419)       0:03:02.145 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating projects] **********************",
                                "Thursday 26 March 2026  03:09:06 +0000 (0:00:06.757)       0:03:08.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating users] *************************",
                                "Thursday 26 March 2026  03:09:06 +0000 (0:00:00.134)       0:03:09.036 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating roles] *************************",
                                "Thursday 26 March 2026  03:09:06 +0000 (0:00:00.137)       0:03:09.174 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Granting user roles] ********************",
                                "Thursday 26 March 2026  03:09:06 +0000 (0:00:00.139)       0:03:09.313 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating default user role] ***********************************",
                                "Thursday 26 March 2026  03:09:06 +0000 (0:00:00.557)       0:03:09.871 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:09:10 +0000 (0:00:03.281)       0:03:13.153 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:09:10 +0000 (0:00:00.512)       0:03:13.665 ******** ",
                                "=============================================================================== ",
                                "keystone : Restart keystone-ssh container ------------------------------ 49.11s",
                                "service-ks-register : keystone | Creating services --------------------- 24.42s",
                                "keystone : Running Keystone bootstrap container ------------------------ 15.23s",
                                "keystone : Creating admin project, user, role, service, and endpoint --- 11.93s",
                                "keystone : Running Keystone fernet bootstrap container ----------------- 11.13s",
                                "keystone : Restart keystone-fernet container --------------------------- 10.44s",
                                "keystone : Copying files for keystone-fernet ---------------------------- 9.15s",
                                "keystone : Restart keystone container ----------------------------------- 7.40s",
                                "service-ks-register : keystone | Creating endpoints --------------------- 6.76s",
                                "keystone : Copying over keystone.conf ----------------------------------- 4.72s",
                                "service-cert-copy : keystone | Copying over extra CA certificates ------- 3.48s",
                                "keystone : Copying over config.json files for services ------------------ 3.32s",
                                "keystone : Creating default user role ----------------------------------- 3.28s",
                                "keystone : Copying files for keystone-ssh ------------------------------- 2.67s",
                                "keystone : Creating keystone database ----------------------------------- 2.36s",
                                "keystone : Check keystone containers ------------------------------------ 2.25s",
                                "keystone : Creating Keystone database user and setting permissions ------ 2.23s",
                                "keystone : Copying over wsgi-keystone.conf ------------------------------ 1.86s",
                                "keystone : Run key distribution ----------------------------------------- 1.78s",
                                "keystone : Ensuring config directories exist ---------------------------- 1.62s",
                                "\u001b[32m2026-03-26 03:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fb909599-55bc-4ffe-8cce-71be6ebf1872 (placement) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fb909599-55bc-4ffe-8cce-71be6ebf1872 (placement) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:09:18 +0000 (0:00:00.270)       0:00:00.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:09:18 +0000 (0:00:00.316)       0:00:00.587 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m",
                                "",
                                "PLAY [Apply role placement] ****************************************************",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:09:19 +0000 (0:00:00.474)       0:00:01.061 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating services] *********************",
                                "Thursday 26 March 2026  03:09:19 +0000 (0:00:00.581)       0:00:01.642 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating endpoints] ********************",
                                "Thursday 26 March 2026  03:09:23 +0000 (0:00:03.627)       0:00:05.269 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating projects] *********************",
                                "Thursday 26 March 2026  03:09:29 +0000 (0:00:06.439)       0:00:11.709 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating users] ************************",
                                "Thursday 26 March 2026  03:09:33 +0000 (0:00:03.547)       0:00:15.256 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating roles] ************************",
                                "Thursday 26 March 2026  03:09:37 +0000 (0:00:04.251)       0:00:19.507 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Granting user roles] *******************",
                                "Thursday 26 March 2026  03:09:40 +0000 (0:00:03.104)       0:00:22.612 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:09:44 +0000 (0:00:04.224)       0:00:26.837 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  03:09:45 +0000 (0:00:00.312)       0:00:27.150 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Check if policies shall be overwritten] **********************",
                                "Thursday 26 March 2026  03:09:46 +0000 (0:00:01.155)       0:00:28.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Set placement policy file] ***********************************",
                                "Thursday 26 March 2026  03:09:46 +0000 (0:00:00.371)       0:00:28.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:09:47 +0000 (0:00:00.350)       0:00:29.028 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over extra CA certificates] ******",
                                "Thursday 26 March 2026  03:09:47 +0000 (0:00:00.583)       0:00:29.611 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:09:49 +0000 (0:00:01.759)       0:00:31.371 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  03:09:50 +0000 (0:00:00.604)       0:00:31.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over config.json files for services] *****************",
                                "Thursday 26 March 2026  03:09:50 +0000 (0:00:00.753)       0:00:32.729 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement.conf] *********************************",
                                "Thursday 26 March 2026  03:09:52 +0000 (0:00:01.693)       0:00:34.423 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement-api wsgi configuration] ***************",
                                "Thursday 26 March 2026  03:09:54 +0000 (0:00:02.350)       0:00:36.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [placement : Copying over migrate-db.rc.j2 configuration] *****************",
                                "Thursday 26 March 2026  03:09:56 +0000 (0:00:01.519)       0:00:38.293 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over existing policy file] ***************************",
                                "Thursday 26 March 2026  03:09:57 +0000 (0:00:01.311)       0:00:39.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Check placement containers] **********************************",
                                "Thursday 26 March 2026  03:09:58 +0000 (0:00:00.803)       0:00:40.408 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases] ********************************",
                                "Thursday 26 March 2026  03:09:59 +0000 (0:00:01.119)       0:00:41.528 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases user and setting permissions] ***",
                                "Thursday 26 March 2026  03:10:01 +0000 (0:00:02.081)       0:00:43.609 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Running placement bootstrap container] ***********************",
                                "Thursday 26 March 2026  03:10:04 +0000 (0:00:02.366)       0:00:45.976 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:10:18 +0000 (0:00:14.182)       0:01:00.158 ******** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:10:18 +0000 (0:00:00.076)       0:01:00.234 ******** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:10:18 +0000 (0:00:00.074)       0:01:00.308 ******** ",
                                "",
                                "RUNNING HANDLER [placement : Restart placement-api container] ******************",
                                "Thursday 26 March 2026  03:10:18 +0000 (0:00:00.071)       0:01:00.380 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:10:28 +0000 (0:00:10.449)       0:01:10.829 ******** ",
                                "=============================================================================== ",
                                "placement : Running placement bootstrap container ---------------------- 14.18s",
                                "placement : Restart placement-api container ---------------------------- 10.45s",
                                "service-ks-register : placement | Creating endpoints -------------------- 6.44s",
                                "service-ks-register : placement | Creating users ------------------------ 4.25s",
                                "service-ks-register : placement | Granting user roles ------------------- 4.22s",
                                "service-ks-register : placement | Creating services --------------------- 3.63s",
                                "service-ks-register : placement | Creating projects --------------------- 3.55s",
                                "service-ks-register : placement | Creating roles ------------------------ 3.10s",
                                "placement : Creating placement databases user and setting permissions --- 2.37s",
                                "placement : Copying over placement.conf --------------------------------- 2.35s",
                                "placement : Creating placement databases -------------------------------- 2.08s",
                                "service-cert-copy : placement | Copying over extra CA certificates ------ 1.76s",
                                "placement : Copying over config.json files for services ----------------- 1.69s",
                                "placement : Copying over placement-api wsgi configuration --------------- 1.52s",
                                "placement : Copying over migrate-db.rc.j2 configuration ----------------- 1.31s",
                                "placement : Ensuring config directories exist --------------------------- 1.16s",
                                "placement : Check placement containers ---------------------------------- 1.12s",
                                "placement : Copying over existing policy file --------------------------- 0.80s",
                                "service-cert-copy : placement | Copying over backend internal TLS key --- 0.75s",
                                "service-cert-copy : placement | Copying over backend internal TLS certificate --- 0.60s",
                                "\u001b[32m2026-03-26 03:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0cd80ac4-a396-4ee1-bd88-7ddd0a9d3192 (neutron) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0cd80ac4-a396-4ee1-bd88-7ddd0a9d3192 (neutron) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:10:36 +0000 (0:00:00.296)       0:00:00.296 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:10:37 +0000 (0:00:00.730)       0:00:01.027 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m",
                                "",
                                "PLAY [Apply role neutron] ******************************************************",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:10:37 +0000 (0:00:00.670)       0:00:01.697 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [neutron : Get container facts] *******************************************",
                                "Thursday 26 March 2026  03:10:39 +0000 (0:00:01.487)       0:00:03.185 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Get container volume facts] ************************************",
                                "Thursday 26 March 2026  03:10:40 +0000 (0:00:01.369)       0:00:04.554 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVN presence] ************************************",
                                "Thursday 26 March 2026  03:10:41 +0000 (0:00:01.138)       0:00:05.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVS presence] ************************************",
                                "Thursday 26 March 2026  03:10:42 +0000 (0:00:00.915)       0:00:06.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating services] ***********************",
                                "Thursday 26 March 2026  03:10:43 +0000 (0:00:00.645)       0:00:07.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating endpoints] **********************",
                                "Thursday 26 March 2026  03:10:47 +0000 (0:00:03.757)       0:00:11.012 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating projects] ***********************",
                                "Thursday 26 March 2026  03:10:53 +0000 (0:00:06.386)       0:00:17.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating users] **************************",
                                "Thursday 26 March 2026  03:10:56 +0000 (0:00:03.132)       0:00:20.531 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating roles] **************************",
                                "Thursday 26 March 2026  03:11:00 +0000 (0:00:03.842)       0:00:24.373 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Granting user roles] *********************",
                                "Thursday 26 March 2026  03:11:03 +0000 (0:00:03.090)       0:00:27.464 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:11:11 +0000 (0:00:08.310)       0:00:35.774 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Load and persist kernel modules] *****************************************",
                                "Thursday 26 March 2026  03:11:12 +0000 (0:00:00.839)       0:00:36.613 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check IPv6 support] ********************************************",
                                "Thursday 26 March 2026  03:11:14 +0000 (0:00:02.199)       0:00:38.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Thursday 26 March 2026  03:11:16 +0000 (0:00:01.211)       0:00:40.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  03:11:18 +0000 (0:00:02.297)       0:00:42.322 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Check if extra ml2 plugins exists] *****************************",
                                "Thursday 26 March 2026  03:11:21 +0000 (0:00:02.642)       0:00:44.965 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m",
                                "\u001b[1;35ma directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:11:21 +0000 (0:00:00.854)       0:00:45.819 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  03:11:23 +0000 (0:00:01.359)       0:00:47.179 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:11:26 +0000 (0:00:03.511)       0:00:50.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  03:11:28 +0000 (0:00:02.225)       0:00:52.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating TLS backend PEM File] *********************************",
                                "Thursday 26 March 2026  03:11:32 +0000 (0:00:03.054)       0:00:55.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:11:34 +0000 (0:00:02.514)       0:00:58.485 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Set neutron policy file] ***************************************",
                                "Thursday 26 March 2026  03:11:34 +0000 (0:00:00.171)       0:00:58.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over existing policy file] *****************************",
                                "Thursday 26 March 2026  03:11:35 +0000 (0:00:00.663)       0:00:59.319 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  03:11:37 +0000 (0:00:02.568)       0:01:01.888 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron.conf] *************************************",
                                "Thursday 26 March 2026  03:11:41 +0000 (0:00:03.393)       0:01:05.281 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_vpnaas.conf] ******************************",
                                "Thursday 26 March 2026  03:11:47 +0000 (0:00:06.438)       0:01:11.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ssh key] ******************************************",
                                "Thursday 26 March 2026  03:11:49 +0000 (0:00:02.044)       0:01:13.764 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ml2_conf.ini] *************************************",
                                "Thursday 26 March 2026  03:11:52 +0000 (0:00:02.729)       0:01:16.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over linuxbridge_agent.ini] ****************************",
                                "Thursday 26 March 2026  03:11:56 +0000 (0:00:03.582)       0:01:20.076 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over openvswitch_agent.ini] ****************************",
                                "Thursday 26 March 2026  03:11:58 +0000 (0:00:02.306)       0:01:22.383 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over sriov_agent.ini] **********************************",
                                "Thursday 26 March 2026  03:12:00 +0000 (0:00:02.333)       0:01:24.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over mlnx_agent.ini] ***********************************",
                                "Thursday 26 March 2026  03:12:03 +0000 (0:00:02.410)       0:01:27.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over eswitchd.conf] ************************************",
                                "Thursday 26 March 2026  03:12:05 +0000 (0:00:02.633)       0:01:29.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dhcp_agent.ini] ***********************************",
                                "Thursday 26 March 2026  03:12:08 +0000 (0:00:02.619)       0:01:32.380 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dnsmasq.conf] *************************************",
                                "Thursday 26 March 2026  03:12:10 +0000 (0:00:02.367)       0:01:34.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over l3_agent.ini] *************************************",
                                "Thursday 26 March 2026  03:12:13 +0000 (0:00:02.958)       0:01:37.706 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over fwaas_driver.ini] *********************************",
                                "Thursday 26 March 2026  03:12:16 +0000 (0:00:02.551)       0:01:40.257 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metadata_agent.ini] *******************************",
                                "Thursday 26 March 2026  03:12:18 +0000 (0:00:02.376)       0:01:42.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************",
                                "Thursday 26 March 2026  03:12:20 +0000 (0:00:02.208)       0:01:44.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metering_agent.ini] *******************************",
                                "Thursday 26 March 2026  03:12:24 +0000 (0:00:03.895)       0:01:48.737 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ironic_neutron_agent.ini] *************************",
                                "Thursday 26 March 2026  03:12:27 +0000 (0:00:02.286)       0:01:51.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over bgp_dragent.ini] **********************************",
                                "Thursday 26 March 2026  03:12:29 +0000 (0:00:02.426)       0:01:53.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ovn_agent.ini] ************************************",
                                "Thursday 26 March 2026  03:12:31 +0000 (0:00:02.371)       0:01:55.822 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over nsx.ini] ******************************************",
                                "Thursday 26 March 2026  03:12:34 +0000 (0:00:02.504)       0:01:58.327 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copy neutron-l3-agent-wrapper script] **************************",
                                "Thursday 26 March 2026  03:12:36 +0000 (0:00:02.256)       0:02:00.583 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over extra ml2 plugins] ********************************",
                                "Thursday 26 March 2026  03:12:39 +0000 (0:00:02.555)       0:02:03.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************",
                                "Thursday 26 March 2026  03:12:41 +0000 (0:00:02.440)       0:02:05.579 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_taas.conf] ********************************",
                                "Thursday 26 March 2026  03:12:43 +0000 (0:00:01.931)       0:02:07.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check neutron containers] **************************************",
                                "Thursday 26 March 2026  03:12:46 +0000 (0:00:02.528)       0:02:10.039 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:12:48 +0000 (0:00:02.731)       0:02:12.770 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database] *************************************",
                                "Thursday 26 March 2026  03:12:49 +0000 (0:00:00.822)       0:02:13.593 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database user and setting permissions] ********",
                                "Thursday 26 March 2026  03:12:51 +0000 (0:00:02.097)       0:02:15.690 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Running Neutron bootstrap container] ***************************",
                                "Thursday 26 March 2026  03:12:54 +0000 (0:00:02.247)       0:02:17.938 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:43.340)       0:03:01.278 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.093)       0:03:01.372 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.073)       0:03:01.446 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.074)       0:03:01.520 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.069)       0:03:01.590 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.070)       0:03:01.661 ******** ",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-server container] *******************",
                                "Thursday 26 March 2026  03:13:37 +0000 (0:00:00.072)       0:03:01.733 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******",
                                "Thursday 26 March 2026  03:14:07 +0000 (0:00:30.095)       0:03:31.829 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:15:10 +0000 (0:01:02.526)       0:04:34.356 ******** ",
                                "=============================================================================== ",
                                "neutron : Restart neutron-ovn-metadata-agent container ----------------- 62.53s",
                                "neutron : Running Neutron bootstrap container -------------------------- 43.34s",
                                "neutron : Restart neutron-server container ----------------------------- 30.10s",
                                "service-ks-register : neutron | Granting user roles --------------------- 8.31s",
                                "neutron : Copying over neutron.conf ------------------------------------- 6.44s",
                                "service-ks-register : neutron | Creating endpoints ---------------------- 6.39s",
                                "neutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 3.90s",
                                "service-ks-register : neutron | Creating users -------------------------- 3.84s",
                                "service-ks-register : neutron | Creating services ----------------------- 3.76s",
                                "neutron : Copying over ml2_conf.ini ------------------------------------- 3.58s",
                                "service-cert-copy : neutron | Copying over extra CA certificates -------- 3.51s",
                                "neutron : Copying over config.json files for services ------------------- 3.39s",
                                "service-ks-register : neutron | Creating projects ----------------------- 3.13s",
                                "service-ks-register : neutron | Creating roles -------------------------- 3.09s",
                                "service-cert-copy : neutron | Copying over backend internal TLS key ----- 3.05s",
                                "neutron : Copying over dnsmasq.conf ------------------------------------- 2.96s",
                                "neutron : Check neutron containers -------------------------------------- 2.73s",
                                "neutron : Copying over ssh key ------------------------------------------ 2.73s",
                                "neutron : Ensuring config directories exist ----------------------------- 2.64s",
                                "neutron : Copying over mlnx_agent.ini ----------------------------------- 2.63s",
                                "\u001b[32m2026-03-26 03:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bd11feb7-7d46-47e4-ae00-c4b2147a1148 (nova) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bd11feb7-7d46-47e4-ae00-c4b2147a1148 (nova) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on OpenStack release] **********************************",
                                "Thursday 26 March 2026  03:15:18 +0000 (0:00:00.306)       0:00:00.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:15:19 +0000 (0:00:00.906)       0:00:01.213 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:15:20 +0000 (0:00:00.958)       0:00:02.172 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova API databases] ********************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Thursday 26 March 2026  03:15:20 +0000 (0:00:00.860)       0:00:03.033 ******** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases] ******************************************",
                                "Thursday 26 March 2026  03:15:21 +0000 (0:00:00.851)       0:00:03.884 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases user and setting permissions] *************",
                                "Thursday 26 March 2026  03:15:25 +0000 (0:00:04.046)       0:00:07.930 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Thursday 26 March 2026  03:15:29 +0000 (0:00:04.101)       0:00:12.032 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for nova-api-bootstrap] ************",
                                "Thursday 26 March 2026  03:15:30 +0000 (0:00:00.681)       0:00:12.714 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************",
                                "Thursday 26 March 2026  03:15:31 +0000 (0:00:01.307)       0:00:14.021 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:15:34 +0000 (0:00:02.696)       0:00:16.718 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Thursday 26 March 2026  03:15:34 +0000 (0:00:00.330)       0:00:17.048 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Create cell0 mappings] ********************************************",
                                "Thursday 26 March 2026  03:16:07 +0000 (0:00:32.739)       0:00:49.788 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Thursday 26 March 2026  03:16:22 +0000 (0:00:14.467)       0:01:04.255 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Thursday 26 March 2026  03:16:33 +0000 (0:00:11.499)       0:01:15.754 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Update cell0 mappings] ********************************************",
                                "Thursday 26 March 2026  03:16:34 +0000 (0:00:00.705)       0:01:16.459 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:16:34 +0000 (0:00:00.511)       0:01:16.970 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Thursday 26 March 2026  03:16:35 +0000 (0:00:00.753)       0:01:17.724 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Thursday 26 March 2026  03:16:52 +0000 (0:00:17.279)       0:01:35.004 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova cell databases] *******************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Thursday 26 March 2026  03:16:53 +0000 (0:00:00.364)       0:01:35.368 ******** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database] *********************************",
                                "Thursday 26 March 2026  03:16:54 +0000 (0:00:00.879)       0:01:36.248 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database user and setting permissions] ****",
                                "Thursday 26 March 2026  03:16:56 +0000 (0:00:02.001)       0:01:38.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Thursday 26 March 2026  03:16:58 +0000 (0:00:02.058)       0:01:40.309 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Thursday 26 March 2026  03:16:58 +0000 (0:00:00.579)       0:01:40.888 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Thursday 26 March 2026  03:17:06 +0000 (0:00:07.383)       0:01:48.271 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Thursday 26 March 2026  03:17:06 +0000 (0:00:00.415)       0:01:48.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  03:17:08 +0000 (0:00:01.576)       0:01:50.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******",
                                "Thursday 26 March 2026  03:17:08 +0000 (0:00:00.527)       0:01:50.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************",
                                "Thursday 26 March 2026  03:17:09 +0000 (0:00:01.006)       0:01:51.796 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Running Nova cell bootstrap container] ***********************",
                                "Thursday 26 March 2026  03:17:12 +0000 (0:00:02.636)       0:01:54.433 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Thursday 26 March 2026  03:17:34 +0000 (0:00:21.755)       0:02:16.188 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Thursday 26 March 2026  03:17:46 +0000 (0:00:12.347)       0:02:28.535 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Create cell] *************************************************",
                                "Thursday 26 March 2026  03:17:47 +0000 (0:00:01.198)       0:02:29.734 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Update cell] *************************************************",
                                "Thursday 26 March 2026  03:18:00 +0000 (0:00:12.613)       0:02:42.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Thursday 26 March 2026  03:18:01 +0000 (0:00:01.116)       0:02:43.463 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova] *********************************************************",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:18:01 +0000 (0:00:00.339)       0:02:43.803 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating services] **************************",
                                "Thursday 26 March 2026  03:18:02 +0000 (0:00:00.809)       0:02:44.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating endpoints] *************************",
                                "Thursday 26 March 2026  03:18:05 +0000 (0:00:03.184)       0:02:47.797 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating projects] **************************",
                                "Thursday 26 March 2026  03:18:11 +0000 (0:00:06.199)       0:02:53.997 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating users] *****************************",
                                "Thursday 26 March 2026  03:18:15 +0000 (0:00:03.242)       0:02:57.240 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating roles] *****************************",
                                "Thursday 26 March 2026  03:18:18 +0000 (0:00:03.849)       0:03:01.089 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Granting user roles] ************************",
                                "Thursday 26 March 2026  03:18:22 +0000 (0:00:03.153)       0:03:04.243 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Thursday 26 March 2026  03:18:29 +0000 (0:00:07.345)       0:03:11.588 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Check if policies shall be overwritten] ***************************",
                                "Thursday 26 March 2026  03:18:30 +0000 (0:00:01.320)       0:03:12.909 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Set nova policy file] *********************************************",
                                "Thursday 26 March 2026  03:18:30 +0000 (0:00:00.157)       0:03:13.067 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check for vendordata file] ****************************************",
                                "Thursday 26 March 2026  03:18:31 +0000 (0:00:00.341)       0:03:13.408 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova : Set vendordata file path] *****************************************",
                                "Thursday 26 March 2026  03:18:31 +0000 (0:00:00.688)       0:03:14.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:18:32 +0000 (0:00:00.570)       0:03:14.668 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Thursday 26 March 2026  03:18:33 +0000 (0:00:00.595)       0:03:15.263 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:18:35 +0000 (0:00:02.389)       0:03:17.653 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Thursday 26 March 2026  03:18:36 +0000 (0:00:00.900)       0:03:18.554 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for services] **********************",
                                "Thursday 26 March 2026  03:18:37 +0000 (0:00:00.859)       0:03:19.413 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf] *******************************************",
                                "Thursday 26 March 2026  03:18:39 +0000 (0:00:02.478)       0:03:21.892 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over existing policy file] ********************************",
                                "Thursday 26 March 2026  03:18:45 +0000 (0:00:06.206)       0:03:28.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-api-wsgi.conf] **********************************",
                                "Thursday 26 March 2026  03:18:46 +0000 (0:00:00.622)       0:03:28.721 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over vendordata file] *************************************",
                                "Thursday 26 March 2026  03:18:48 +0000 (0:00:01.647)       0:03:30.368 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check nova containers] ********************************************",
                                "Thursday 26 March 2026  03:18:48 +0000 (0:00:00.342)       0:03:30.710 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.2.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:18:50 +0000 (0:00:02.115)       0:03:32.826 ******** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:18:50 +0000 (0:00:00.150)       0:03:32.977 ******** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:18:50 +0000 (0:00:00.139)       0:03:33.117 ******** ",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-scheduler container] **********************",
                                "Thursday 26 March 2026  03:18:51 +0000 (0:00:00.148)       0:03:33.265 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-api container] ****************************",
                                "Thursday 26 March 2026  03:19:15 +0000 (0:00:23.876)       0:03:57.142 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role nova-cell] ****************************************************",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:19:20 +0000 (0:00:05.424)       0:04:02.566 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:19:21 +0000 (0:00:01.330)       0:04:03.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Load and persist br_netfilter module] ************************************",
                                "Thursday 26 March 2026  03:19:22 +0000 (0:00:00.950)       0:04:04.847 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Thursday 26 March 2026  03:19:23 +0000 (0:00:00.895)       0:04:05.742 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Thursday 26 March 2026  03:19:24 +0000 (0:00:00.866)       0:04:06.609 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Thursday 26 March 2026  03:19:25 +0000 (0:00:01.247)       0:04:07.856 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************",
                                "Thursday 26 March 2026  03:19:26 +0000 (0:00:00.604)       0:04:08.461 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "",
                                "TASK [nova-cell : Install udev kolla kvm rules] ********************************",
                                "Thursday 26 March 2026  03:19:28 +0000 (0:00:02.015)       0:04:10.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Mask qemu-kvm service] ***************************************",
                                "Thursday 26 March 2026  03:19:29 +0000 (0:00:01.311)       0:04:11.789 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  03:19:31 +0000 (0:00:01.830)       0:04:13.619 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:19:33 +0000 (0:00:02.391)       0:04:16.011 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Thursday 26 March 2026  03:19:35 +0000 (0:00:01.483)       0:04:17.495 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:19:39 +0000 (0:00:03.709)       0:04:21.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Thursday 26 March 2026  03:19:40 +0000 (0:00:01.779)       0:04:22.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:19:43 +0000 (0:00:02.222)       0:04:25.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova keyring file] *************************************",
                                "Thursday 26 March 2026  03:19:44 +0000 (0:00:01.124)       0:04:26.330 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check cinder keyring file] ***********************************",
                                "Thursday 26 March 2026  03:19:45 +0000 (0:00:01.156)       0:04:27.487 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract nova key from file] **********************************",
                                "Thursday 26 March 2026  03:19:46 +0000 (0:00:00.973)       0:04:28.461 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract cinder key from file] ********************************",
                                "Thursday 26 March 2026  03:19:46 +0000 (0:00:00.558)       0:04:29.020 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph nova keyring file] ****************************",
                                "Thursday 26 March 2026  03:19:47 +0000 (0:00:00.540)       0:04:29.560 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph cinder keyring file] **************************",
                                "Thursday 26 March 2026  03:19:48 +0000 (0:00:01.434)       0:04:30.995 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf] *****************************************",
                                "Thursday 26 March 2026  03:19:50 +0000 (0:00:01.238)       0:04:32.233 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************",
                                "Thursday 26 March 2026  03:19:54 +0000 (0:00:03.905)       0:04:36.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************",
                                "Thursday 26 March 2026  03:19:54 +0000 (0:00:00.346)       0:04:36.484 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring libvirt secrets directory exists] *******************",
                                "Thursday 26 March 2026  03:19:54 +0000 (0:00:00.555)       0:04:37.040 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing nova secret xml for libvirt] *************************",
                                "Thursday 26 March 2026  03:19:56 +0000 (0:00:01.266)       0:04:38.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing secrets key for libvirt] *****************************",
                                "Thursday 26 March 2026  03:19:59 +0000 (0:00:03.341)       0:04:41.648 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check if policies shall be overwritten] **********************",
                                "Thursday 26 March 2026  03:20:02 +0000 (0:00:03.453)       0:04:45.101 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set nova policy file] ****************************************",
                                "Thursday 26 March 2026  03:20:03 +0000 (0:00:00.149)       0:04:45.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check for vendordata file] ***********************************",
                                "Thursday 26 March 2026  03:20:04 +0000 (0:00:00.932)       0:04:46.183 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set vendordata file path] ************************************",
                                "Thursday 26 March 2026  03:20:04 +0000 (0:00:00.737)       0:04:46.921 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for services] *****************",
                                "Thursday 26 March 2026  03:20:05 +0000 (0:00:00.852)       0:04:47.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf] **************************************",
                                "Thursday 26 March 2026  03:20:09 +0000 (0:00:03.846)       0:04:51.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over Nova compute provider config] *******************",
                                "Thursday 26 March 2026  03:20:16 +0000 (0:00:06.982)       0:04:58.601 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt configuration] **************************",
                                "Thursday 26 March 2026  03:20:17 +0000 (0:00:01.365)       0:04:59.966 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt TLS keys] *******************************",
                                "Thursday 26 March 2026  03:20:21 +0000 (0:00:03.720)       0:05:03.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt SASL configuration] *********************",
                                "Thursday 26 March 2026  03:20:22 +0000 (0:00:00.637)       0:05:04.324 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying files for nova-ssh] **********************************",
                                "Thursday 26 March 2026  03:20:27 +0000 (0:00:05.383)       0:05:09.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying VMware vCenter CA file] ******************************",
                                "Thursday 26 March 2026  03:20:35 +0000 (0:00:07.669)       0:05:17.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying 'release' file for nova_compute] *********************",
                                "Thursday 26 March 2026  03:20:36 +0000 (0:00:00.839)       0:05:18.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************",
                                "Thursday 26 March 2026  03:20:36 +0000 (0:00:00.654)       0:05:18.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over existing policy file] ***************************",
                                "Thursday 26 March 2026  03:20:38 +0000 (0:00:02.172)       0:05:21.043 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over vendordata file to containers] ******************",
                                "Thursday 26 March 2026  03:20:40 +0000 (0:00:01.563)       0:05:22.607 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova-cell containers] **********************************",
                                "Thursday 26 March 2026  03:20:41 +0000 (0:00:00.968)       0:05:23.576 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:20:44 +0000 (0:00:02.706)       0:05:26.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:44 +0000 (0:00:00.830)       0:05:27.113 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:45 +0000 (0:00:00.169)       0:05:27.282 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:45 +0000 (0:00:00.149)       0:05:27.432 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:45 +0000 (0:00:00.154)       0:05:27.587 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:45 +0000 (0:00:00.148)       0:05:27.735 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:20:45 +0000 (0:00:00.324)       0:05:28.060 ******** ",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************",
                                "Thursday 26 March 2026  03:20:46 +0000 (0:00:00.157)       0:05:28.218 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************",
                                "Thursday 26 March 2026  03:20:53 +0000 (0:00:07.475)       0:05:35.693 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************",
                                "Thursday 26 March 2026  03:21:09 +0000 (0:00:15.667)       0:05:51.361 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************",
                                "Thursday 26 March 2026  03:21:36 +0000 (0:00:27.431)       0:06:18.792 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************",
                                "Thursday 26 March 2026  03:22:19 +0000 (0:00:43.096)       0:07:01.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************",
                                "Thursday 26 March 2026  03:22:20 +0000 (0:00:00.751)       0:07:02.640 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-compute container] *******************",
                                "Thursday 26 March 2026  03:22:21 +0000 (0:00:00.848)       0:07:03.489 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***",
                                "Thursday 26 March 2026  03:22:47 +0000 (0:00:25.699)       0:07:29.189 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Waiting for nova-compute services to register themselves] ****",
                                "Thursday 26 March 2026  03:22:47 +0000 (0:00:00.148)       0:07:29.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if nova-compute service failed to register] *************",
                                "Thursday 26 March 2026  03:23:10 +0000 (0:00:22.999)       0:07:52.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include discover_computes.yml] *******************************",
                                "Thursday 26 March 2026  03:23:20 +0000 (0:00:10.321)       0:08:02.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Thursday 26 March 2026  03:23:25 +0000 (0:00:04.471)       0:08:07.130 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Thursday 26 March 2026  03:23:37 +0000 (0:00:12.681)       0:08:19.811 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if cell settings not found] *****************************",
                                "Thursday 26 March 2026  03:23:39 +0000 (0:00:01.569)       0:08:21.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Discover nova hosts] *****************************************",
                                "Thursday 26 March 2026  03:23:41 +0000 (0:00:01.801)       0:08:23.182 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************",
                                "Thursday 26 March 2026  03:23:52 +0000 (0:00:11.180)       0:08:34.362 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Refresh nova scheduler cell cache] ***************************************",
                                "",
                                "TASK [nova : Refresh cell cache in nova scheduler] *****************************",
                                "Thursday 26 March 2026  03:23:54 +0000 (0:00:01.936)       0:08:36.298 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova super conductor services] *****************************",
                                "",
                                "TASK [nova : Reload nova super conductor services to remove RPC version pin] ***",
                                "Thursday 26 March 2026  03:23:55 +0000 (0:00:00.954)       0:08:37.253 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload Nova cell services] ***********************************************",
                                "",
                                "TASK [nova-cell : Reload nova cell services to remove RPC version cap] *********",
                                "Thursday 26 March 2026  03:23:55 +0000 (0:00:00.770)       0:08:38.023 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova API services] *****************************************",
                                "",
                                "TASK [nova : Reload nova API services to remove RPC version pin] ***************",
                                "Thursday 26 March 2026  03:23:57 +0000 (0:00:01.448)       0:08:39.472 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Run Nova API online data migrations] *************************************",
                                "",
                                "TASK [nova : Run Nova API online database migrations] **************************",
                                "Thursday 26 March 2026  03:23:58 +0000 (0:00:00.647)       0:08:40.119 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Run Nova cell online data migrations] ************************************",
                                "",
                                "TASK [nova-cell : Run Nova cell online database migrations] ********************",
                                "Thursday 26 March 2026  03:23:58 +0000 (0:00:00.986)       0:08:41.106 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=44  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:23:59 +0000 (0:00:00.495)       0:08:41.601 ******** ",
                                "=============================================================================== ",
                                "nova-cell : Restart nova-libvirt container ----------------------------- 43.10s",
                                "nova : Running Nova API bootstrap container ---------------------------- 32.74s",
                                "nova-cell : Restart nova-ssh container --------------------------------- 27.43s",
                                "nova-cell : Restart nova-compute container ----------------------------- 25.70s",
                                "nova : Restart nova-scheduler container -------------------------------- 23.88s",
                                "nova-cell : Waiting for nova-compute services to register themselves --- 23.00s",
                                "nova-cell : Running Nova cell bootstrap container ---------------------- 21.76s",
                                "nova : Running Nova API bootstrap container ---------------------------- 17.28s",
                                "nova-cell : Restart nova-novncproxy container -------------------------- 15.67s",
                                "nova : Create cell0 mappings ------------------------------------------- 14.47s",
                                "nova-cell : Get a list of existing cells ------------------------------- 12.68s",
                                "nova-cell : Create cell ------------------------------------------------ 12.61s",
                                "nova-cell : Get a list of existing cells ------------------------------- 12.35s",
                                "nova-cell : Get a list of existing cells ------------------------------- 11.50s",
                                "nova-cell : Discover nova hosts ---------------------------------------- 11.18s",
                                "nova-cell : Fail if nova-compute service failed to register ------------ 10.32s",
                                "nova-cell : Copying files for nova-ssh ---------------------------------- 7.67s",
                                "nova-cell : Restart nova-conductor container ---------------------------- 7.48s",
                                "service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 7.38s",
                                "service-ks-register : nova | Granting user roles ------------------------ 7.35s",
                                "\u001b[32m2026-03-26 03:24:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f721f98e-b0b6-4aeb-a88b-3394e455017a (horizon) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:24:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f721f98e-b0b6-4aeb-a88b-3394e455017a (horizon) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:24:06 +0000 (0:00:00.301)       0:00:00.301 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:24:07 +0000 (0:00:00.356)       0:00:00.657 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m",
                                "",
                                "PLAY [Apply role horizon] ******************************************************",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:24:07 +0000 (0:00:00.454)       0:00:01.112 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  03:24:08 +0000 (0:00:00.535)       0:00:01.647 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : Set empty custom policy] ***************************************",
                                "Thursday 26 March 2026  03:24:09 +0000 (0:00:01.204)       0:00:02.852 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:24:09 +0000 (0:00:00.538)       0:00:03.390 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:10 +0000 (0:00:00.825)       0:00:04.216 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:11 +0000 (0:00:00.371)       0:00:04.588 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:11 +0000 (0:00:00.330)       0:00:04.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:11 +0000 (0:00:00.325)       0:00:05.244 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:12 +0000 (0:00:00.344)       0:00:05.589 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:12 +0000 (0:00:00.149)       0:00:05.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:12 +0000 (0:00:00.341)       0:00:06.080 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:13 +0000 (0:00:00.540)       0:00:06.620 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:13 +0000 (0:00:00.134)       0:00:06.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:13 +0000 (0:00:00.332)       0:00:07.087 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:13 +0000 (0:00:00.325)       0:00:07.413 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:14 +0000 (0:00:00.148)       0:00:07.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:14 +0000 (0:00:00.568)       0:00:08.131 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:15 +0000 (0:00:00.343)       0:00:08.475 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:15 +0000 (0:00:00.143)       0:00:08.618 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:15 +0000 (0:00:00.335)       0:00:08.953 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:15 +0000 (0:00:00.350)       0:00:09.303 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:16 +0000 (0:00:00.334)       0:00:09.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:16 +0000 (0:00:00.358)       0:00:09.997 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:16 +0000 (0:00:00.353)       0:00:10.350 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:17 +0000 (0:00:00.165)       0:00:10.516 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:17 +0000 (0:00:00.323)       0:00:10.840 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:17 +0000 (0:00:00.587)       0:00:11.427 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:18 +0000 (0:00:00.153)       0:00:11.581 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:18 +0000 (0:00:00.374)       0:00:11.955 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:18 +0000 (0:00:00.395)       0:00:12.351 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:19 +0000 (0:00:00.129)       0:00:12.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Thursday 26 March 2026  03:24:19 +0000 (0:00:00.620)       0:00:13.100 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:24:20 +0000 (0:00:00.389)       0:00:13.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Thursday 26 March 2026  03:24:20 +0000 (0:00:00.156)       0:00:13.647 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  03:24:20 +0000 (0:00:00.340)       0:00:13.987 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over horizon.conf] *************************************",
                                "Thursday 26 March 2026  03:24:22 +0000 (0:00:02.095)       0:00:16.082 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over kolla-settings.py] ********************************",
                                "Thursday 26 March 2026  03:24:24 +0000 (0:00:02.088)       0:00:18.171 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom-settings.py] *******************************",
                                "Thursday 26 March 2026  03:24:26 +0000 (0:00:01.882)       0:00:20.053 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over existing policy file] *****************************",
                                "Thursday 26 March 2026  03:24:28 +0000 (0:00:01.562)       0:00:21.616 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom themes] ************************************",
                                "Thursday 26 March 2026  03:24:28 +0000 (0:00:00.566)       0:00:22.182 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:24:29 +0000 (0:00:00.294)       0:00:22.477 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  03:24:29 +0000 (0:00:00.652)       0:00:23.129 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:24:31 +0000 (0:00:02.119)       0:00:25.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  03:24:32 +0000 (0:00:00.667)       0:00:25.917 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Deploy horizon container] **************************************",
                                "Thursday 26 March 2026  03:24:33 +0000 (0:00:00.894)       0:00:26.811 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.2.20251130', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:24:35 +0000 (0:00:01.794)       0:00:28.606 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:24:35 +0000 (0:00:00.325)       0:00:28.931 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database] *************************************",
                                "Thursday 26 March 2026  03:24:36 +0000 (0:00:00.600)       0:00:29.532 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database user and setting permissions] ********",
                                "Thursday 26 March 2026  03:24:38 +0000 (0:00:02.347)       0:00:31.879 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Running Horizon bootstrap container] ***************************",
                                "Thursday 26 March 2026  03:24:41 +0000 (0:00:02.704)       0:00:34.583 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:24:58 +0000 (0:00:17.350)       0:00:51.934 ******** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:24:58 +0000 (0:00:00.087)       0:00:52.022 ******** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:24:58 +0000 (0:00:00.066)       0:00:52.089 ******** ",
                                "",
                                "RUNNING HANDLER [horizon : Restart horizon container] **************************",
                                "Thursday 26 March 2026  03:24:58 +0000 (0:00:00.082)       0:00:52.171 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:25:21 +0000 (0:00:22.557)       0:01:14.729 ******** ",
                                "=============================================================================== ",
                                "horizon : Restart horizon container ------------------------------------ 22.56s",
                                "horizon : Running Horizon bootstrap container -------------------------- 17.35s",
                                "horizon : Creating Horizon database user and setting permissions -------- 2.70s",
                                "horizon : Creating Horizon database ------------------------------------- 2.35s",
                                "service-cert-copy : horizon | Copying over extra CA certificates -------- 2.12s",
                                "horizon : Copying over config.json files for services ------------------- 2.10s",
                                "horizon : Copying over horizon.conf ------------------------------------- 2.09s",
                                "horizon : Copying over kolla-settings.py -------------------------------- 1.88s",
                                "horizon : Deploy horizon container -------------------------------------- 1.79s",
                                "horizon : Copying over custom-settings.py ------------------------------- 1.56s",
                                "horizon : Ensuring config directories exist ----------------------------- 1.20s",
                                "service-cert-copy : horizon | Copying over backend internal TLS key ----- 0.89s",
                                "horizon : include_tasks ------------------------------------------------- 0.83s",
                                "service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.67s",
                                "horizon : include_tasks ------------------------------------------------- 0.65s",
                                "horizon : Update custom policy file name -------------------------------- 0.62s",
                                "horizon : include_tasks ------------------------------------------------- 0.60s",
                                "horizon : Update policy file name --------------------------------------- 0.59s",
                                "horizon : Update custom policy file name -------------------------------- 0.57s",
                                "horizon : Copying over existing policy file ----------------------------- 0.57s",
                                "\u001b[32m2026-03-26 03:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 845c9a2d-6656-4f8a-8038-0ee897175f20 (skyline) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 845c9a2d-6656-4f8a-8038-0ee897175f20 (skyline) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:25:28 +0000 (0:00:00.270)       0:00:00.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:25:28 +0000 (0:00:00.313)       0:00:00.584 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_skyline_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_skyline_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_skyline_True)\u001b[0m",
                                "",
                                "PLAY [Apply role skyline] ******************************************************",
                                "",
                                "TASK [skyline : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:25:29 +0000 (0:00:00.463)       0:00:01.048 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/skyline/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Creating services] ***********************",
                                "Thursday 26 March 2026  03:25:29 +0000 (0:00:00.589)       0:00:01.637 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=skyline (panel))\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Creating endpoints] **********************",
                                "Thursday 26 March 2026  03:25:33 +0000 (0:00:03.454)       0:00:05.092 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> https://api-int.testbed.osism.xyz:9998 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> https://api.testbed.osism.xyz:9998 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Creating projects] ***********************",
                                "Thursday 26 March 2026  03:25:40 +0000 (0:00:06.814)       0:00:11.906 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Creating users] **************************",
                                "Thursday 26 March 2026  03:25:43 +0000 (0:00:02.984)       0:00:14.891 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Creating roles] **************************",
                                "Thursday 26 March 2026  03:25:47 +0000 (0:00:04.064)       0:00:18.955 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : skyline | Granting user roles] *********************",
                                "Thursday 26 March 2026  03:25:50 +0000 (0:00:03.407)       0:00:22.363 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=skyline -> service -> admin)\u001b[0m",
                                "",
                                "TASK [skyline : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  03:25:54 +0000 (0:00:03.844)       0:00:26.207 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [skyline : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:25:55 +0000 (0:00:01.339)       0:00:27.547 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/skyline/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : skyline | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  03:25:56 +0000 (0:00:00.777)       0:00:28.324 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : skyline | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:25:59 +0000 (0:00:02.460)       0:00:30.785 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : skyline | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  03:25:59 +0000 (0:00:00.790)       0:00:31.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [skyline : Copying over skyline.yaml files for services] ******************",
                                "Thursday 26 March 2026  03:26:01 +0000 (0:00:01.340)       0:00:32.917 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [skyline : Copying over gunicorn.py files for services] *******************",
                                "Thursday 26 March 2026  03:26:03 +0000 (0:00:02.498)       0:00:35.415 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/skyline/templates/gunicorn.py.j2)\u001b[0m",
                                "",
                                "TASK [skyline : Copying over nginx.conf files for services] ********************",
                                "Thursday 26 March 2026  03:26:05 +0000 (0:00:01.639)       0:00:37.055 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/skyline/templates/nginx.conf.j2)\u001b[0m",
                                "",
                                "TASK [skyline : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  03:26:07 +0000 (0:00:02.137)       0:00:39.193 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [skyline : Copying over custom logos] *************************************",
                                "Thursday 26 March 2026  03:26:09 +0000 (0:00:02.307)       0:00:41.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [skyline : Check skyline container] ***************************************",
                                "Thursday 26 March 2026  03:26:10 +0000 (0:00:00.325)       0:00:41.826 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [skyline : Creating Skyline database] *************************************",
                                "Thursday 26 March 2026  03:26:12 +0000 (0:00:01.836)       0:00:43.663 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [skyline : Creating Skyline database user and setting permissions] ********",
                                "Thursday 26 March 2026  03:26:14 +0000 (0:00:02.288)       0:00:45.952 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [skyline : Running Skyline bootstrap container] ***************************",
                                "Thursday 26 March 2026  03:26:16 +0000 (0:00:02.406)       0:00:48.358 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [skyline : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:26:24 +0000 (0:00:08.143)       0:00:56.502 ******** ",
                                "",
                                "TASK [skyline : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:26:24 +0000 (0:00:00.074)       0:00:56.576 ******** ",
                                "",
                                "TASK [skyline : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:26:25 +0000 (0:00:00.071)       0:00:56.648 ******** ",
                                "",
                                "RUNNING HANDLER [skyline : Restart skyline-apiserver container] ****************",
                                "Thursday 26 March 2026  03:26:25 +0000 (0:00:00.083)       0:00:56.731 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [skyline : Restart skyline-console container] ******************",
                                "Thursday 26 March 2026  03:26:36 +0000 (0:00:11.142)       0:01:07.874 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:26:46 +0000 (0:00:09.775)       0:01:17.650 ******** ",
                                "=============================================================================== ",
                                "skyline : Restart skyline-apiserver container -------------------------- 11.14s",
                                "skyline : Restart skyline-console container ----------------------------- 9.78s",
                                "skyline : Running Skyline bootstrap container --------------------------- 8.14s",
                                "service-ks-register : skyline | Creating endpoints ---------------------- 6.81s",
                                "service-ks-register : skyline | Creating users -------------------------- 4.06s",
                                "service-ks-register : skyline | Granting user roles --------------------- 3.84s",
                                "service-ks-register : skyline | Creating services ----------------------- 3.45s",
                                "service-ks-register : skyline | Creating roles -------------------------- 3.41s",
                                "service-ks-register : skyline | Creating projects ----------------------- 2.98s",
                                "skyline : Copying over skyline.yaml files for services ------------------ 2.50s",
                                "service-cert-copy : skyline | Copying over extra CA certificates -------- 2.46s",
                                "skyline : Creating Skyline database user and setting permissions -------- 2.41s",
                                "skyline : Copying over config.json files for services ------------------- 2.31s",
                                "skyline : Creating Skyline database ------------------------------------- 2.29s",
                                "skyline : Copying over nginx.conf files for services -------------------- 2.14s",
                                "skyline : Check skyline container --------------------------------------- 1.84s",
                                "skyline : Copying over gunicorn.py files for services ------------------- 1.64s",
                                "service-cert-copy : skyline | Copying over backend internal TLS key ----- 1.34s",
                                "skyline : Ensuring config directories exist ----------------------------- 1.34s",
                                "service-cert-copy : skyline | Copying over backend internal TLS certificate --- 0.79s",
                                "\u001b[32m2026-03-26 03:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06ad8f27-7892-4279-9b53-bb6de8a558f5 (glance) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06ad8f27-7892-4279-9b53-bb6de8a558f5 (glance) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:26:53 +0000 (0:00:00.298)       0:00:00.298 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:26:53 +0000 (0:00:00.436)       0:00:00.735 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m",
                                "",
                                "PLAY [Apply role glance] *******************************************************",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:26:54 +0000 (0:00:00.489)       0:00:01.225 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating services] ************************",
                                "Thursday 26 March 2026  03:26:54 +0000 (0:00:00.676)       0:00:01.901 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating endpoints] ***********************",
                                "Thursday 26 March 2026  03:26:58 +0000 (0:00:03.400)       0:00:05.301 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating projects] ************************",
                                "Thursday 26 March 2026  03:27:04 +0000 (0:00:06.406)       0:00:11.708 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating users] ***************************",
                                "Thursday 26 March 2026  03:27:07 +0000 (0:00:03.290)       0:00:14.999 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating roles] ***************************",
                                "Thursday 26 March 2026  03:27:12 +0000 (0:00:04.347)       0:00:19.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Granting user roles] **********************",
                                "Thursday 26 March 2026  03:27:15 +0000 (0:00:03.609)       0:00:22.956 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  03:27:20 +0000 (0:00:04.187)       0:00:27.144 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:27:23 +0000 (0:00:03.578)       0:00:30.723 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [glance : Ensuring glance service ceph config subdir exists] **************",
                                "Thursday 26 March 2026  03:27:24 +0000 (0:00:00.796)       0:00:31.519 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copy over multiple ceph configs for Glance] *********************",
                                "Thursday 26 March 2026  03:27:28 +0000 (0:00:03.690)       0:00:35.210 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Copy over ceph Glance keyrings] *********************************",
                                "Thursday 26 March 2026  03:27:29 +0000 (0:00:01.659)       0:00:36.869 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directory has correct owner and permission] *****",
                                "Thursday 26 March 2026  03:27:31 +0000 (0:00:01.448)       0:00:38.318 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check if policies shall be overwritten] *************************",
                                "Thursday 26 March 2026  03:27:32 +0000 (0:00:00.736)       0:00:39.055 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Set glance policy file] *****************************************",
                                "Thursday 26 March 2026  03:27:32 +0000 (0:00:00.139)       0:00:39.195 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:27:32 +0000 (0:00:00.336)       0:00:39.531 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  03:27:33 +0000 (0:00:00.910)       0:00:40.441 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:27:37 +0000 (0:00:04.121)       0:00:44.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  03:27:40 +0000 (0:00:03.221)       0:00:47.784 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating TLS backend PEM File] **********************************",
                                "Thursday 26 March 2026  03:27:44 +0000 (0:00:03.615)       0:00:51.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  03:27:47 +0000 (0:00:03.243)       0:00:54.643 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-api.conf] ***********************************",
                                "Thursday 26 March 2026  03:27:51 +0000 (0:00:04.066)       0:00:58.710 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-cache.conf for glance_api] ******************",
                                "Thursday 26 March 2026  03:27:57 +0000 (0:00:05.749)       0:01:04.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-swift.conf for glance_api] ******************",
                                "Thursday 26 March 2026  03:28:01 +0000 (0:00:03.616)       0:01:08.077 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-image-import.conf] **************************",
                                "Thursday 26 March 2026  03:28:04 +0000 (0:00:03.784)       0:01:11.861 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over property-protections-rules.conf] *******************",
                                "Thursday 26 March 2026  03:28:08 +0000 (0:00:03.647)       0:01:15.509 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over existing policy file] ******************************",
                                "Thursday 26 March 2026  03:28:12 +0000 (0:00:03.667)       0:01:19.177 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-haproxy-tls.cfg] ****************************",
                                "Thursday 26 March 2026  03:28:12 +0000 (0:00:00.582)       0:01:19.759 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostnqn' file for glance_api] ***********************",
                                "Thursday 26 March 2026  03:28:16 +0000 (0:00:03.340)       0:01:23.100 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check glance containers] ****************************************",
                                "Thursday 26 March 2026  03:28:20 +0000 (0:00:04.405)       0:01:27.506 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:28:24 +0000 (0:00:03.983)       0:01:31.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database] ***************************************",
                                "Thursday 26 March 2026  03:28:25 +0000 (0:00:00.565)       0:01:32.056 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database user and setting permissions] **********",
                                "Thursday 26 March 2026  03:28:27 +0000 (0:00:02.210)       0:01:34.267 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Enable log_bin_trust_function_creators function] ****************",
                                "Thursday 26 March 2026  03:28:29 +0000 (0:00:02.446)       0:01:36.714 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Running Glance bootstrap container] *****************************",
                                "Thursday 26 March 2026  03:28:32 +0000 (0:00:02.423)       0:01:39.137 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Disable log_bin_trust_function_creators function] ***************",
                                "Thursday 26 March 2026  03:29:00 +0000 (0:00:28.855)       0:02:07.993 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:29:03 +0000 (0:00:02.219)       0:02:10.213 ******** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:29:03 +0000 (0:00:00.070)       0:02:10.283 ******** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:29:03 +0000 (0:00:00.071)       0:02:10.355 ******** ",
                                "",
                                "RUNNING HANDLER [glance : Restart glance-api container] ************************",
                                "Thursday 26 March 2026  03:29:03 +0000 (0:00:00.076)       0:02:10.432 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:29:37 +0000 (0:00:33.781)       0:02:44.214 ******** ",
                                "=============================================================================== ",
                                "glance : Restart glance-api container ---------------------------------- 33.78s",
                                "glance : Running Glance bootstrap container ---------------------------- 28.86s",
                                "service-ks-register : glance | Creating endpoints ----------------------- 6.41s",
                                "glance : Copying over glance-api.conf ----------------------------------- 5.75s",
                                "glance : Generating 'hostnqn' file for glance_api ----------------------- 4.41s",
                                "service-ks-register : glance | Creating users --------------------------- 4.35s",
                                "service-ks-register : glance | Granting user roles ---------------------- 4.19s",
                                "service-cert-copy : glance | Copying over extra CA certificates --------- 4.12s",
                                "glance : Copying over config.json files for services -------------------- 4.07s",
                                "glance : Check glance containers ---------------------------------------- 3.98s",
                                "glance : Copying over glance-swift.conf for glance_api ------------------ 3.78s",
                                "glance : Ensuring glance service ceph config subdir exists -------------- 3.69s",
                                "glance : Copying over property-protections-rules.conf ------------------- 3.67s",
                                "glance : Copying over glance-image-import.conf -------------------------- 3.65s",
                                "glance : Copying over glance-cache.conf for glance_api ------------------ 3.62s",
                                "service-cert-copy : glance | Copying over backend internal TLS key ------ 3.62s",
                                "service-ks-register : glance | Creating roles --------------------------- 3.61s",
                                "glance : Ensuring config directories exist ------------------------------ 3.58s",
                                "service-ks-register : glance | Creating services ------------------------ 3.40s",
                                "glance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.34s",
                                "\u001b[32m2026-03-26 03:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08eb036f-e85f-4744-96a6-01108fe8d3ed (cinder) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08eb036f-e85f-4744-96a6-01108fe8d3ed (cinder) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:29:44 +0000 (0:00:00.269)       0:00:00.269 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:29:44 +0000 (0:00:00.342)       0:00:00.612 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m",
                                "",
                                "PLAY [Apply role cinder] *******************************************************",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:29:45 +0000 (0:00:00.465)       0:00:01.077 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating services] ************************",
                                "Thursday 26 March 2026  03:29:45 +0000 (0:00:00.589)       0:00:01.667 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating endpoints] ***********************",
                                "Thursday 26 March 2026  03:29:49 +0000 (0:00:03.433)       0:00:05.100 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating projects] ************************",
                                "Thursday 26 March 2026  03:29:55 +0000 (0:00:06.462)       0:00:11.563 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating users] ***************************",
                                "Thursday 26 March 2026  03:29:58 +0000 (0:00:03.230)       0:00:14.793 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating roles] ***************************",
                                "Thursday 26 March 2026  03:30:02 +0000 (0:00:04.060)       0:00:18.853 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Granting user roles] **********************",
                                "Thursday 26 March 2026  03:30:06 +0000 (0:00:03.240)       0:00:22.094 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  03:30:13 +0000 (0:00:07.226)       0:00:29.321 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:30:15 +0000 (0:00:02.050)       0:00:31.371 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:30:15 +0000 (0:00:00.520)       0:00:31.892 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring cinder service ceph config subdirs exists] *************",
                                "Thursday 26 March 2026  03:30:16 +0000 (0:00:00.544)       0:00:32.437 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over multiple ceph.conf for cinder services] ************",
                                "Thursday 26 March 2026  03:30:18 +0000 (0:00:01.624)       0:00:34.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************",
                                "Thursday 26 March 2026  03:30:21 +0000 (0:00:03.500)       0:00:37.561 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************",
                                "Thursday 26 March 2026  03:30:23 +0000 (0:00:01.630)       0:00:39.192 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directory has correct owner and permission] *****",
                                "Thursday 26 March 2026  03:30:25 +0000 (0:00:02.668)       0:00:41.860 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Check if policies shall be overwritten] *************************",
                                "Thursday 26 March 2026  03:30:26 +0000 (0:00:01.061)       0:00:42.921 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Set cinder policy file] *****************************************",
                                "Thursday 26 March 2026  03:30:27 +0000 (0:00:00.167)       0:00:43.089 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:30:27 +0000 (0:00:00.510)       0:00:43.599 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  03:30:28 +0000 (0:00:00.604)       0:00:44.204 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:30:32 +0000 (0:00:04.200)       0:00:48.405 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  03:30:33 +0000 (0:00:00.986)       0:00:49.391 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  03:30:34 +0000 (0:00:00.966)       0:00:50.357 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder-wsgi.conf] **********************************",
                                "Thursday 26 March 2026  03:30:38 +0000 (0:00:04.517)       0:00:54.875 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder.conf] ***************************************",
                                "Thursday 26 March 2026  03:30:40 +0000 (0:00:01.957)       0:00:56.833 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************",
                                "Thursday 26 March 2026  03:30:52 +0000 (0:00:11.264)       0:01:08.098 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over existing policy file] ******************************",
                                "Thursday 26 March 2026  03:30:53 +0000 (0:00:01.566)       0:01:09.664 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over nfs_shares files for cinder_volume] ****************",
                                "Thursday 26 March 2026  03:30:54 +0000 (0:00:00.950)       0:01:10.615 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Check cinder containers] ****************************************",
                                "Thursday 26 March 2026  03:30:55 +0000 (0:00:00.618)       0:01:11.233 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:30:58 +0000 (0:00:02.975)       0:01:14.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database] ***************************************",
                                "Thursday 26 March 2026  03:30:58 +0000 (0:00:00.307)       0:01:14.516 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database user and setting permissions] **********",
                                "Thursday 26 March 2026  03:31:00 +0000 (0:00:02.141)       0:01:16.657 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Running Cinder bootstrap container] *****************************",
                                "Thursday 26 March 2026  03:31:02 +0000 (0:00:02.182)       0:01:18.840 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:31:23 +0000 (0:00:20.199)       0:01:39.039 ******** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:31:23 +0000 (0:00:00.070)       0:01:39.110 ******** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:31:23 +0000 (0:00:00.076)       0:01:39.187 ******** ",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-api container] ************************",
                                "Thursday 26 March 2026  03:31:23 +0000 (0:00:00.074)       0:01:39.261 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************",
                                "Thursday 26 March 2026  03:31:49 +0000 (0:00:26.597)       0:02:05.859 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-volume container] *********************",
                                "Thursday 26 March 2026  03:31:58 +0000 (0:00:08.377)       0:02:14.237 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-backup container] *********************",
                                "Thursday 26 March 2026  03:32:25 +0000 (0:00:27.727)       0:02:41.964 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***",
                                "Thursday 26 March 2026  03:32:37 +0000 (0:00:11.022)       0:02:52.986 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:32:37 +0000 (0:00:00.284)       0:02:53.271 ******** ",
                                "=============================================================================== ",
                                "cinder : Restart cinder-volume container ------------------------------- 27.73s",
                                "cinder : Restart cinder-api container ---------------------------------- 26.60s",
                                "cinder : Running Cinder bootstrap container ---------------------------- 20.20s",
                                "cinder : Copying over cinder.conf -------------------------------------- 11.26s",
                                "cinder : Restart cinder-backup container ------------------------------- 11.02s",
                                "cinder : Restart cinder-scheduler container ----------------------------- 8.38s",
                                "service-ks-register : cinder | Granting user roles ---------------------- 7.23s",
                                "service-ks-register : cinder | Creating endpoints ----------------------- 6.46s",
                                "cinder : Copying over config.json files for services -------------------- 4.52s",
                                "service-cert-copy : cinder | Copying over extra CA certificates --------- 4.20s",
                                "service-ks-register : cinder | Creating users --------------------------- 4.06s",
                                "cinder : Copying over multiple ceph.conf for cinder services ------------ 3.50s",
                                "service-ks-register : cinder | Creating services ------------------------ 3.43s",
                                "service-ks-register : cinder | Creating roles --------------------------- 3.24s",
                                "service-ks-register : cinder | Creating projects ------------------------ 3.23s",
                                "cinder : Check cinder containers ---------------------------------------- 2.98s",
                                "cinder : Copy over Ceph keyring files for cinder-backup ----------------- 2.67s",
                                "cinder : Creating Cinder database user and setting permissions ---------- 2.18s",
                                "cinder : Creating Cinder database --------------------------------------- 2.14s",
                                "cinder : Ensuring config directories exist ------------------------------ 2.05s",
                                "\u001b[32m2026-03-26 03:32:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 68ea24c0-1a9f-4835-83c3-be9f58b464fd (barbican) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:32:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 68ea24c0-1a9f-4835-83c3-be9f58b464fd (barbican) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:32:44 +0000 (0:00:00.286)       0:00:00.286 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:32:44 +0000 (0:00:00.332)       0:00:00.618 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m",
                                "",
                                "PLAY [Apply role barbican] *****************************************************",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:32:45 +0000 (0:00:00.496)       0:00:01.114 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating services] **********************",
                                "Thursday 26 March 2026  03:32:45 +0000 (0:00:00.582)       0:00:01.697 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating endpoints] *********************",
                                "Thursday 26 March 2026  03:32:49 +0000 (0:00:03.493)       0:00:05.190 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating projects] **********************",
                                "Thursday 26 March 2026  03:32:55 +0000 (0:00:06.389)       0:00:11.580 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating users] *************************",
                                "Thursday 26 March 2026  03:32:59 +0000 (0:00:03.211)       0:00:14.791 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating roles] *************************",
                                "Thursday 26 March 2026  03:33:03 +0000 (0:00:04.371)       0:00:19.162 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Granting user roles] ********************",
                                "Thursday 26 March 2026  03:33:19 +0000 (0:00:16.079)       0:00:35.242 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring config directories exist] ****************************",
                                "Thursday 26 March 2026  03:33:23 +0000 (0:00:03.806)       0:00:39.049 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring vassals config directories exist] ********************",
                                "Thursday 26 March 2026  03:33:24 +0000 (0:00:01.624)       0:00:40.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m",
                                "",
                                "TASK [barbican : Check if policies shall be overwritten] ***********************",
                                "Thursday 26 March 2026  03:33:26 +0000 (0:00:01.236)       0:00:41.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Set barbican policy file] *************************************",
                                "Thursday 26 March 2026  03:33:26 +0000 (0:00:00.368)       0:00:42.278 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:33:26 +0000 (0:00:00.326)       0:00:42.604 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over extra CA certificates] *******",
                                "Thursday 26 March 2026  03:33:27 +0000 (0:00:00.581)       0:00:43.185 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:33:30 +0000 (0:00:03.472)       0:00:46.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****",
                                "Thursday 26 March 2026  03:33:31 +0000 (0:00:00.584)       0:00:47.242 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over config.json files for services] ******************",
                                "Thursday 26 March 2026  03:33:32 +0000 (0:00:00.828)       0:00:48.070 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api.ini] ********************************",
                                "Thursday 26 March 2026  03:33:35 +0000 (0:00:03.616)       0:00:51.687 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Checking whether barbican-api-paste.ini file exists] **********",
                                "Thursday 26 March 2026  03:33:37 +0000 (0:00:01.625)       0:00:53.313 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api-paste.ini] **************************",
                                "Thursday 26 March 2026  03:33:38 +0000 (0:00:01.031)       0:00:54.345 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican.conf] ***********************************",
                                "Thursday 26 March 2026  03:33:39 +0000 (0:00:00.592)       0:00:54.938 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over existing policy file] ****************************",
                                "Thursday 26 March 2026  03:33:45 +0000 (0:00:06.794)       0:01:01.732 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Check barbican containers] ************************************",
                                "Thursday 26 March 2026  03:33:46 +0000 (0:00:00.856)       0:01:02.588 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:33:49 +0000 (0:00:02.397)       0:01:04.986 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database] ***********************************",
                                "Thursday 26 March 2026  03:33:49 +0000 (0:00:00.328)       0:01:05.314 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database user and setting permissions] ******",
                                "Thursday 26 March 2026  03:33:51 +0000 (0:00:02.031)       0:01:07.345 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Running barbican bootstrap container] *************************",
                                "Thursday 26 March 2026  03:33:53 +0000 (0:00:02.220)       0:01:09.566 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:34:06 +0000 (0:00:12.503)       0:01:22.069 ******** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:34:06 +0000 (0:00:00.073)       0:01:22.142 ******** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Thursday 26 March 2026  03:34:06 +0000 (0:00:00.083)       0:01:22.226 ******** ",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-api container] ********************",
                                "Thursday 26 March 2026  03:34:06 +0000 (0:00:00.082)       0:01:22.309 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******",
                                "Thursday 26 March 2026  03:34:18 +0000 (0:00:11.526)       0:01:33.835 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-worker container] *****************",
                                "Thursday 26 March 2026  03:34:23 +0000 (0:00:05.071)       0:01:38.907 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:34:33 +0000 (0:00:10.439)       0:01:49.346 ******** ",
                                "=============================================================================== ",
                                "service-ks-register : barbican | Creating roles ------------------------ 16.08s",
                                "barbican : Running barbican bootstrap container ------------------------ 12.50s",
                                "barbican : Restart barbican-api container ------------------------------ 11.53s",
                                "barbican : Restart barbican-worker container --------------------------- 10.44s",
                                "barbican : Copying over barbican.conf ----------------------------------- 6.79s",
                                "service-ks-register : barbican | Creating endpoints --------------------- 6.39s",
                                "barbican : Restart barbican-keystone-listener container ----------------- 5.07s",
                                "service-ks-register : barbican | Creating users ------------------------- 4.37s",
                                "service-ks-register : barbican | Granting user roles -------------------- 3.81s",
                                "barbican : Copying over config.json files for services ------------------ 3.62s",
                                "service-ks-register : barbican | Creating services ---------------------- 3.49s",
                                "service-cert-copy : barbican | Copying over extra CA certificates ------- 3.47s",
                                "service-ks-register : barbican | Creating projects ---------------------- 3.21s",
                                "barbican : Check barbican containers ------------------------------------ 2.40s",
                                "barbican : Creating barbican database user and setting permissions ------ 2.22s",
                                "barbican : Creating barbican database ----------------------------------- 2.03s",
                                "barbican : Copying over barbican-api.ini -------------------------------- 1.63s",
                                "barbican : Ensuring config directories exist ---------------------------- 1.62s",
                                "barbican : Ensuring vassals config directories exist -------------------- 1.24s",
                                "barbican : Checking whether barbican-api-paste.ini file exists ---------- 1.03s",
                                "\u001b[32m2026-03-26 03:34:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9fd4a647-2541-4c2c-a7fe-e47d95f2c814 (designate) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:34:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9fd4a647-2541-4c2c-a7fe-e47d95f2c814 (designate) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:34:40 +0000 (0:00:00.295)       0:00:00.295 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:34:41 +0000 (0:00:00.332)       0:00:00.628 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m",
                                "",
                                "PLAY [Apply role designate] ****************************************************",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:34:41 +0000 (0:00:00.485)       0:00:01.113 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating services] *********************",
                                "Thursday 26 March 2026  03:34:42 +0000 (0:00:00.570)       0:00:01.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating endpoints] ********************",
                                "Thursday 26 March 2026  03:34:45 +0000 (0:00:03.465)       0:00:05.149 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating projects] *********************",
                                "Thursday 26 March 2026  03:34:52 +0000 (0:00:06.464)       0:00:11.614 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating users] ************************",
                                "Thursday 26 March 2026  03:34:55 +0000 (0:00:03.235)       0:00:14.850 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating roles] ************************",
                                "Thursday 26 March 2026  03:34:59 +0000 (0:00:04.052)       0:00:18.902 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Granting user roles] *******************",
                                "Thursday 26 March 2026  03:35:02 +0000 (0:00:03.286)       0:00:22.188 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m",
                                "",
                                "TASK [designate : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  03:35:06 +0000 (0:00:03.969)       0:00:26.158 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Check if policies shall be overwritten] **********************",
                                "Thursday 26 March 2026  03:35:09 +0000 (0:00:02.998)       0:00:29.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Set designate policy file] ***********************************",
                                "Thursday 26 March 2026  03:35:10 +0000 (0:00:00.146)       0:00:29.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:35:10 +0000 (0:00:00.584)       0:00:29.888 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over extra CA certificates] ******",
                                "Thursday 26 March 2026  03:35:11 +0000 (0:00:00.574)       0:00:30.462 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:35:17 +0000 (0:00:06.039)       0:00:36.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  03:35:18 +0000 (0:00:01.058)       0:00:37.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [designate : Copying over config.json files for services] *****************",
                                "Thursday 26 March 2026  03:35:19 +0000 (0:00:01.214)       0:00:38.774 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over designate.conf] *********************************",
                                "Thursday 26 March 2026  03:35:25 +0000 (0:00:06.202)       0:00:44.976 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over pools.yaml] *************************************",
                                "Thursday 26 March 2026  03:35:40 +0000 (0:00:15.070)       0:01:00.046 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over named.conf] *************************************",
                                "Thursday 26 March 2026  03:35:44 +0000 (0:00:03.756)       0:01:03.802 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.conf] **************************************",
                                "Thursday 26 March 2026  03:35:47 +0000 (0:00:02.632)       0:01:06.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.key] ***************************************",
                                "Thursday 26 March 2026  03:35:50 +0000 (0:00:03.033)       0:01:09.469 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:35:53 +0000 (0:00:02.845)       0:01:12.314 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over existing policy file] ***************************",
                                "Thursday 26 March 2026  03:35:53 +0000 (0:00:00.331)       0:01:12.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Check designate containers] **********************************",
                                "Thursday 26 March 2026  03:35:54 +0000 (0:00:00.800)       0:01:13.447 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  03:35:58 +0000 (0:00:04.719)       0:01:18.166 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases] ********************************",
                                "Thursday 26 March 2026  03:35:59 +0000 (0:00:00.339)       0:01:18.506 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases user and setting permissions] ***",
                                "Thursday 26 March 2026  03:36:01 +0000 (0:00:02.148)       0:01:20.655 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m",
                                "",
                                "TASK [designate : Running Designate bootstrap container] ***********************",
                                "Thursday 26 March 2026  03:36:03 +0000 (0:00:02.342)       0:01:22.997 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:36:19 +0000 (0:00:16.271)       0:01:39.269 ******** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:36:20 +0000 (0:00:00.073)       0:01:39.342 ******** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Thursday 26 March 2026  03:36:20 +0000 (0:00:00.074)       0:01:39.417 ******** ",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********",
                                "Thursday 26 March 2026  03:36:20 +0000 (0:00:00.072)       0:01:39.489 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-api container] ******************",
                                "Thursday 26 March 2026  03:36:28 +0000 (0:00:07.943)       0:01:47.432 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-central container] **************",
                                "Thursday 26 March 2026  03:36:38 +0000 (0:00:10.718)       0:01:58.151 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-producer container] *************",
                                "Thursday 26 March 2026  03:36:44 +0000 (0:00:05.939)       0:02:04.091 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-mdns container] *****************",
                                "Thursday 26 March 2026  03:36:50 +0000 (0:00:05.941)       0:02:10.032 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-worker container] ***************",
                                "Thursday 26 March 2026  03:36:56 +0000 (0:00:06.071)       0:02:16.104 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [designate : Non-destructive DNS pools update] ****************************",
                                "Thursday 26 March 2026  03:37:07 +0000 (0:00:11.022)       0:02:27.127 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:37:15 +0000 (0:00:07.390)       0:02:34.517 ******** ",
                                "=============================================================================== ",
                                "designate : Running Designate bootstrap container ---------------------- 16.27s",
                                "designate : Copying over designate.conf -------------------------------- 15.07s",
                                "designate : Restart designate-worker container ------------------------- 11.02s",
                                "designate : Restart designate-api container ---------------------------- 10.72s",
                                "designate : Restart designate-backend-bind9 container ------------------- 7.94s",
                                "designate : Non-destructive DNS pools update ---------------------------- 7.39s",
                                "service-ks-register : designate | Creating endpoints -------------------- 6.46s",
                                "designate : Copying over config.json files for services ----------------- 6.20s",
                                "designate : Restart designate-mdns container ---------------------------- 6.07s",
                                "service-cert-copy : designate | Copying over extra CA certificates ------ 6.04s",
                                "designate : Restart designate-producer container ------------------------ 5.94s",
                                "designate : Restart designate-central container ------------------------- 5.94s",
                                "designate : Check designate containers ---------------------------------- 4.72s",
                                "service-ks-register : designate | Creating users ------------------------ 4.05s",
                                "service-ks-register : designate | Granting user roles ------------------- 3.97s",
                                "designate : Copying over pools.yaml ------------------------------------- 3.76s",
                                "service-ks-register : designate | Creating services --------------------- 3.47s",
                                "service-ks-register : designate | Creating roles ------------------------ 3.29s",
                                "service-ks-register : designate | Creating projects --------------------- 3.24s",
                                "designate : Copying over rndc.conf -------------------------------------- 3.03s",
                                "\u001b[32m2026-03-26 03:37:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 435223f9-b996-4394-82a5-4bcb72d6599b (octavia) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:37:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 435223f9-b996-4394-82a5-4bcb72d6599b (octavia) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:37:22 +0000 (0:00:00.271)       0:00:00.271 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:37:23 +0000 (0:00:00.307)       0:00:00.579 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m",
                                "",
                                "PLAY [Apply role octavia] ******************************************************",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:37:23 +0000 (0:00:00.461)       0:00:01.041 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating services] ***********************",
                                "Thursday 26 March 2026  03:37:24 +0000 (0:00:00.579)       0:00:01.621 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating endpoints] **********************",
                                "Thursday 26 March 2026  03:37:27 +0000 (0:00:03.307)       0:00:04.928 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating projects] ***********************",
                                "Thursday 26 March 2026  03:37:34 +0000 (0:00:06.535)       0:00:11.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating users] **************************",
                                "Thursday 26 March 2026  03:37:37 +0000 (0:00:03.320)       0:00:14.785 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating roles] **************************",
                                "Thursday 26 March 2026  03:37:45 +0000 (0:00:08.315)       0:00:23.100 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Granting user roles] *********************",
                                "Thursday 26 March 2026  03:37:49 +0000 (0:00:03.289)       0:00:26.390 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "",
                                "TASK [octavia : Adding octavia related roles] **********************************",
                                "Thursday 26 March 2026  03:37:56 +0000 (0:00:07.620)       0:00:34.010 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:38:12 +0000 (0:00:16.050)       0:00:50.060 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create amphora flavor] *****************************************",
                                "Thursday 26 March 2026  03:38:13 +0000 (0:00:00.780)       0:00:50.840 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create nova keypair for amphora] *******************************",
                                "Thursday 26 March 2026  03:38:18 +0000 (0:00:05.038)       0:00:55.879 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Thursday 26 March 2026  03:38:23 +0000 (0:00:04.454)       0:01:00.333 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create security groups for octavia] ****************************",
                                "Thursday 26 March 2026  03:38:26 +0000 (0:00:03.281)       0:01:03.615 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Add rules for security groups] *********************************",
                                "Thursday 26 March 2026  03:38:36 +0000 (0:00:10.054)       0:01:13.669 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management network] ************************",
                                "Thursday 26 March 2026  03:38:53 +0000 (0:00:16.922)       0:01:30.592 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management subnet] *************************",
                                "Thursday 26 March 2026  03:38:57 +0000 (0:00:04.607)       0:01:35.199 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management router for IPv6] ****************",
                                "Thursday 26 March 2026  03:39:03 +0000 (0:00:05.523)       0:01:40.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update loadbalancer management subnet] *************************",
                                "Thursday 26 March 2026  03:39:03 +0000 (0:00:00.221)       0:01:40.944 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:39:08 +0000 (0:00:04.459)       0:01:45.403 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create ports for Octavia health-manager nodes] *****************",
                                "Thursday 26 March 2026  03:39:09 +0000 (0:00:01.158)       0:01:46.561 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update Octavia health manager port host_id] ********************",
                                "Thursday 26 March 2026  03:39:14 +0000 (0:00:05.629)       0:01:52.191 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Add Octavia port to openvswitch br-int] ************************",
                                "Thursday 26 March 2026  03:39:19 +0000 (0:00:04.313)       0:01:56.505 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Install isc-dhcp-client package] *******************************",
                                "Thursday 26 March 2026  03:39:20 +0000 (0:00:01.047)       0:01:57.552 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia dhclient conf] **********************************",
                                "Thursday 26 March 2026  03:39:21 +0000 (0:00:01.716)       0:01:59.268 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia-interface service] ******************************",
                                "Thursday 26 March 2026  03:39:23 +0000 (0:00:01.276)       0:02:00.545 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Restart octavia-interface.service if required] *****************",
                                "Thursday 26 March 2026  03:39:24 +0000 (0:00:01.188)       0:02:01.734 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Enable and start octavia-interface.service] ********************",
                                "Thursday 26 March 2026  03:39:26 +0000 (0:00:02.205)       0:02:03.939 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Wait for interface ohm0 ip appear] *****************************",
                                "Thursday 26 March 2026  03:39:28 +0000 (0:00:01.515)       0:02:05.455 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Gather facts] **************************************************",
                                "Thursday 26 March 2026  03:39:28 +0000 (0:00:00.643)       0:02:06.099 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:39:31 +0000 (0:00:03.116)       0:02:09.216 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Get amphora flavor info] ***************************************",
                                "Thursday 26 March 2026  03:39:32 +0000 (0:00:00.575)       0:02:09.791 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Thursday 26 March 2026  03:39:36 +0000 (0:00:03.990)       0:02:13.781 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get security groups for octavia] *******************************",
                                "Thursday 26 March 2026  03:39:39 +0000 (0:00:03.294)       0:02:17.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Get loadbalancer management network] ***************************",
                                "Thursday 26 March 2026  03:39:47 +0000 (0:00:07.349)       0:02:24.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia resources facts] ***********************************",
                                "Thursday 26 March 2026  03:39:51 +0000 (0:00:04.216)       0:02:28.641 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  03:39:51 +0000 (0:00:00.555)       0:02:29.197 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Check if policies shall be overwritten] ************************",
                                "Thursday 26 March 2026  03:39:54 +0000 (0:00:02.522)       0:02:31.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia policy file] ***************************************",
                                "Thursday 26 March 2026  03:39:54 +0000 (0:00:00.140)       0:02:31.860 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over existing policy file] *****************************",
                                "Thursday 26 March 2026  03:39:54 +0000 (0:00:00.311)       0:02:32.171 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:39:55 +0000 (0:00:00.704)       0:02:32.876 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  03:39:56 +0000 (0:00:00.754)       0:02:33.631 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:40:01 +0000 (0:00:04.963)       0:02:38.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  03:40:01 +0000 (0:00:00.668)       0:02:39.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  03:40:03 +0000 (0:00:01.337)       0:02:40.600 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia-wsgi.conf] ********************************",
                                "Thursday 26 March 2026  03:40:08 +0000 (0:00:05.318)       0:02:45.919 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia.conf] *************************************",
                                "Thursday 26 March 2026  03:40:10 +0000 (0:00:01.698)       0:02:47.617 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over Octavia SSH key] **********************************",
                                "Thursday 26 March 2026  03:40:27 +0000 (0:00:17.437)       0:03:05.055 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-worker] ******************",
                                "Thursday 26 March 2026  03:40:29 +0000 (0:00:01.902)       0:03:06.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-housekeeping] ************",
                                "Thursday 26 March 2026  03:40:34 +0000 (0:00:05.103)       0:03:12.061 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-health-manager] **********",
                                "Thursday 26 March 2026  03:40:39 +0000 (0:00:05.205)       0:03:17.266 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Check octavia containers] **************************************",
                                "Thursday 26 March 2026  03:40:45 +0000 (0:00:05.318)       0:03:22.585 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:40:49 +0000 (0:00:04.089)       0:03:26.675 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database] *************************************",
                                "Thursday 26 March 2026  03:40:49 +0000 (0:00:00.588)       0:03:27.263 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database] *************************",
                                "Thursday 26 March 2026  03:40:52 +0000 (0:00:02.237)       0:03:29.501 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database user and setting permissions] ********",
                                "Thursday 26 March 2026  03:40:54 +0000 (0:00:02.128)       0:03:31.630 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database user and setting permissions] ***",
                                "Thursday 26 March 2026  03:40:56 +0000 (0:00:02.233)       0:03:33.863 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Running Octavia bootstrap container] ***************************",
                                "Thursday 26 March 2026  03:40:58 +0000 (0:00:02.189)       0:03:36.053 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:41:21 +0000 (0:00:23.142)       0:03:59.196 ******** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:41:21 +0000 (0:00:00.073)       0:03:59.269 ******** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:41:22 +0000 (0:00:00.070)       0:03:59.340 ******** ",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-api container] **********************",
                                "Thursday 26 March 2026  03:41:22 +0000 (0:00:00.071)       0:03:59.411 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************",
                                "Thursday 26 March 2026  03:41:39 +0000 (0:00:17.243)       0:04:16.654 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********",
                                "Thursday 26 March 2026  03:41:50 +0000 (0:00:11.439)       0:04:28.093 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************",
                                "Thursday 26 March 2026  03:42:01 +0000 (0:00:10.580)       0:04:38.674 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-worker container] *******************",
                                "Thursday 26 March 2026  03:42:11 +0000 (0:00:10.104)       0:04:48.778 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:42:17 +0000 (0:00:05.841)       0:04:54.619 ******** ",
                                "=============================================================================== ",
                                "octavia : Running Octavia bootstrap container -------------------------- 23.14s",
                                "octavia : Copying over octavia.conf ------------------------------------ 17.44s",
                                "octavia : Restart octavia-api container -------------------------------- 17.24s",
                                "octavia : Add rules for security groups -------------------------------- 16.92s",
                                "octavia : Adding octavia related roles --------------------------------- 16.05s",
                                "octavia : Restart octavia-driver-agent container ----------------------- 11.44s",
                                "octavia : Restart octavia-health-manager container --------------------- 10.58s",
                                "octavia : Restart octavia-housekeeping container ----------------------- 10.10s",
                                "octavia : Create security groups for octavia --------------------------- 10.05s",
                                "service-ks-register : octavia | Creating users -------------------------- 8.32s",
                                "service-ks-register : octavia | Granting user roles --------------------- 7.62s",
                                "octavia : Get security groups for octavia ------------------------------- 7.35s",
                                "service-ks-register : octavia | Creating endpoints ---------------------- 6.54s",
                                "octavia : Restart octavia-worker container ------------------------------ 5.84s",
                                "octavia : Create ports for Octavia health-manager nodes ----------------- 5.63s",
                                "octavia : Create loadbalancer management subnet ------------------------- 5.52s",
                                "octavia : Copying certificate files for octavia-health-manager ---------- 5.32s",
                                "octavia : Copying over config.json files for services ------------------- 5.32s",
                                "octavia : Copying certificate files for octavia-housekeeping ------------ 5.21s",
                                "octavia : Copying certificate files for octavia-worker ------------------ 5.10s",
                                "\u001b[32m2026-03-26 03:42:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ced74339-9146-4768-949d-5ed80733d53c (ceilometer) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:42:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ced74339-9146-4768-949d-5ed80733d53c (ceilometer) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:42:24 +0000 (0:00:00.277)       0:00:00.277 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:42:25 +0000 (0:00:00.720)       0:00:00.998 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceilometer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceilometer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceilometer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceilometer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceilometer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceilometer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceilometer] ***************************************************",
                                "",
                                "TASK [ceilometer : include_tasks] **********************************************",
                                "Thursday 26 March 2026  03:42:25 +0000 (0:00:00.636)       0:00:01.635 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceilometer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Creating services] ********************",
                                "Thursday 26 March 2026  03:42:27 +0000 (0:00:01.271)       0:00:02.906 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Creating endpoints] *******************",
                                "Thursday 26 March 2026  03:42:27 +0000 (0:00:00.131)       0:00:03.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Creating projects] ********************",
                                "Thursday 26 March 2026  03:42:27 +0000 (0:00:00.136)       0:00:03.174 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Creating users] ***********************",
                                "Thursday 26 March 2026  03:42:31 +0000 (0:00:03.889)       0:00:07.064 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceilometer -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Creating roles] ***********************",
                                "Thursday 26 March 2026  03:42:35 +0000 (0:00:04.001)       0:00:11.065 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceilometer | Granting user roles] ******************",
                                "Thursday 26 March 2026  03:42:38 +0000 (0:00:03.231)       0:00:14.296 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceilometer -> service -> admin)\u001b[0m",
                                "",
                                "TASK [ceilometer : Associate the ResellerAdmin role and ceilometer user] *******",
                                "Thursday 26 March 2026  03:42:42 +0000 (0:00:04.305)       0:00:18.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceilometer : Ensuring config directories exist] **************************",
                                "Thursday 26 March 2026  03:42:42 +0000 (0:00:00.139)       0:00:18.742 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [ceilometer : Check if the folder for custom meter definitions exist] *****",
                                "Thursday 26 March 2026  03:42:44 +0000 (0:00:01.453)       0:00:20.196 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set variable that indicates if we have a folder for custom meter YAML files] ***",
                                "Thursday 26 March 2026  03:42:46 +0000 (0:00:01.683)       0:00:21.880 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Find all *.yaml files in custom meter definitions folder (if the folder exist)] ***",
                                "Thursday 26 March 2026  03:42:46 +0000 (0:00:00.623)       0:00:22.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set the variable that control the copy of custom meter definitions] ***",
                                "Thursday 26 March 2026  03:42:47 +0000 (0:00:00.861)       0:00:23.364 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Create default folder for custom meter definitions] *********",
                                "Thursday 26 March 2026  03:42:48 +0000 (0:00:00.685)       0:00:24.049 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying custom meter definitions to Ceilometer] *************",
                                "Thursday 26 March 2026  03:42:49 +0000 (0:00:01.070)       0:00:25.120 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check if the folder [\"/opt/configuration/environments/kolla/files/overlays/ceilometer/pollsters.d\"] for dynamic pollsters definitions exist] ***",
                                "Thursday 26 March 2026  03:42:50 +0000 (0:00:00.871)       0:00:25.991 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set the variable that control the copy of dynamic pollsters definitions] ***",
                                "Thursday 26 March 2026  03:42:50 +0000 (0:00:00.712)       0:00:26.703 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Clean default folder for dynamic pollsters definitions] *****",
                                "Thursday 26 March 2026  03:42:51 +0000 (0:00:00.846)       0:00:27.550 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Create default folder for dynamic pollsters definitions] ****",
                                "Thursday 26 March 2026  03:42:52 +0000 (0:00:00.943)       0:00:28.494 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying dynamic pollsters definitions] **********************",
                                "Thursday 26 March 2026  03:42:53 +0000 (0:00:00.835)       0:00:29.329 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check if custom polling.yaml exists] ************************",
                                "Thursday 26 March 2026  03:42:54 +0000 (0:00:00.625)       0:00:29.954 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over polling.yaml] **********************************",
                                "Thursday 26 March 2026  03:42:55 +0000 (0:00:01.632)       0:00:31.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set ceilometer polling file's path] *************************",
                                "Thursday 26 March 2026  03:42:56 +0000 (0:00:00.814)       0:00:32.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check custom gnocchi_resources.yaml exists] *****************",
                                "Thursday 26 March 2026  03:42:57 +0000 (0:00:00.823)       0:00:33.224 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over gnocchi_resources.yaml] ************************",
                                "Thursday 26 March 2026  03:42:58 +0000 (0:00:01.344)       0:00:34.569 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set ceilometer gnocchi_resources file's path] ***************",
                                "Thursday 26 March 2026  03:42:59 +0000 (0:00:01.094)       0:00:35.664 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check if policies shall be overwritten] *********************",
                                "Thursday 26 March 2026  03:43:00 +0000 (0:00:00.877)       0:00:36.542 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceilometer : Set ceilometer policy file] *********************************",
                                "Thursday 26 March 2026  03:43:00 +0000 (0:00:00.148)       0:00:36.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : include_tasks] **********************************************",
                                "Thursday 26 March 2026  03:43:01 +0000 (0:00:00.651)       0:00:37.341 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceilometer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : ceilometer | Copying over extra CA certificates] *****",
                                "Thursday 26 March 2026  03:43:02 +0000 (0:00:01.357)       0:00:38.699 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : ceilometer | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:43:05 +0000 (0:00:02.407)       0:00:41.106 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : ceilometer | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  03:43:06 +0000 (0:00:00.923)       0:00:42.029 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over config.json files for services] ****************",
                                "Thursday 26 March 2026  03:43:07 +0000 (0:00:01.596)       0:00:43.626 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over ceilometer.conf] *******************************",
                                "Thursday 26 March 2026  03:43:10 +0000 (0:00:02.796)       0:00:46.422 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [ceilometer : Check custom event_definitions.yaml exists] *****************",
                                "Thursday 26 March 2026  03:43:15 +0000 (0:00:05.232)       0:00:51.654 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over event_definitions.yaml] ************************",
                                "Thursday 26 March 2026  03:43:17 +0000 (0:00:01.732)       0:00:53.386 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over event_definitions.yaml for notification service] ***",
                                "Thursday 26 March 2026  03:43:18 +0000 (0:00:00.613)       0:00:54.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over event_pipeline.yaml] ***************************",
                                "Thursday 26 March 2026  03:43:19 +0000 (0:00:01.664)       0:00:55.665 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check custom pipeline.yaml exists] **************************",
                                "Thursday 26 March 2026  03:43:21 +0000 (0:00:01.535)       0:00:57.201 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over custom pipeline.yaml file] *********************",
                                "Thursday 26 March 2026  03:43:22 +0000 (0:00:01.620)       0:00:58.821 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over pipeline.yaml file] ****************************",
                                "Thursday 26 March 2026  03:43:25 +0000 (0:00:02.674)       0:01:01.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying VMware vCenter CA file] *****************************",
                                "Thursday 26 March 2026  03:43:26 +0000 (0:00:00.881)       0:01:02.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Copying over existing policy file] **************************",
                                "Thursday 26 March 2026  03:43:27 +0000 (0:00:00.836)       0:01:03.214 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Check ceilometer containers] ********************************",
                                "Thursday 26 March 2026  03:43:28 +0000 (0:00:00.862)       0:01:04.076 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-notification', 'value': {'container_name': 'ceilometer_notification', 'group': 'ceilometer-notification', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-notification/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-agent-notification 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ceilometer-compute', 'value': {'container_name': 'ceilometer_compute', 'group': 'ceilometer-compute', 'enabled': True, 'privileged': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', 'nova_libvirt:/var/lib/libvirt', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ceilometer-polling 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceilometer-central', 'value': {'container_name': 'ceilometer_central', 'group': 'ceilometer-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130', 'volumes': ['/etc/kolla/ceilometer-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'ceilometer:/var/lib/ceilometer/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [ceilometer : include_tasks] **********************************************",
                                "Thursday 26 March 2026  03:43:29 +0000 (0:00:01.655)       0:01:05.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceilometer : Running Ceilometer bootstrap container] *********************",
                                "Thursday 26 March 2026  03:43:30 +0000 (0:00:00.653)       0:01:06.385 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:04.792)       0:01:11.178 ******** ",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:00.073)       0:01:11.251 ******** ",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:00.109)       0:01:11.361 ******** ",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:00.254)       0:01:11.615 ******** ",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:00.071)       0:01:11.686 ******** ",
                                "",
                                "TASK [ceilometer : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:43:35 +0000 (0:00:00.070)       0:01:11.757 ******** ",
                                "",
                                "RUNNING HANDLER [ceilometer : Restart ceilometer-notification container] *******",
                                "Thursday 26 March 2026  03:43:36 +0000 (0:00:00.074)       0:01:11.831 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceilometer : Restart ceilometer-central container] ************",
                                "Thursday 26 March 2026  03:43:43 +0000 (0:00:07.594)       0:01:19.425 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceilometer : Restart ceilometer-compute container] ************",
                                "Thursday 26 March 2026  03:43:48 +0000 (0:00:04.951)       0:01:24.377 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:43:59 +0000 (0:00:11.229)       0:01:35.606 ******** ",
                                "=============================================================================== ",
                                "ceilometer : Restart ceilometer-compute container ---------------------- 11.23s",
                                "ceilometer : Restart ceilometer-notification container ------------------ 7.59s",
                                "ceilometer : Copying over ceilometer.conf ------------------------------- 5.23s",
                                "ceilometer : Restart ceilometer-central container ----------------------- 4.95s",
                                "ceilometer : Running Ceilometer bootstrap container --------------------- 4.79s",
                                "service-ks-register : ceilometer | Granting user roles ------------------ 4.31s",
                                "service-ks-register : ceilometer | Creating users ----------------------- 4.00s",
                                "service-ks-register : ceilometer | Creating projects -------------------- 3.89s",
                                "service-ks-register : ceilometer | Creating roles ----------------------- 3.23s",
                                "ceilometer : Copying over config.json files for services ---------------- 2.80s",
                                "ceilometer : Copying over custom pipeline.yaml file --------------------- 2.67s",
                                "service-cert-copy : ceilometer | Copying over extra CA certificates ----- 2.41s",
                                "ceilometer : Check custom event_definitions.yaml exists ----------------- 1.73s",
                                "ceilometer : Check if the folder for custom meter definitions exist ----- 1.68s",
                                "ceilometer : Copying over event_definitions.yaml for notification service --- 1.66s",
                                "ceilometer : Check ceilometer containers -------------------------------- 1.66s",
                                "ceilometer : Check if custom polling.yaml exists ------------------------ 1.63s",
                                "ceilometer : Check custom pipeline.yaml exists -------------------------- 1.62s",
                                "service-cert-copy : ceilometer | Copying over backend internal TLS key --- 1.60s",
                                "ceilometer : Copying over event_pipeline.yaml --------------------------- 1.54s",
                                "\u001b[32m2026-03-26 03:44:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dff533f4-b9a6-47d4-badc-74d13e9ba245 (aodh) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:44:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dff533f4-b9a6-47d4-badc-74d13e9ba245 (aodh) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:44:07 +0000 (0:00:00.286)       0:00:00.286 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:44:07 +0000 (0:00:00.309)       0:00:00.596 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_aodh_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_aodh_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_aodh_True)\u001b[0m",
                                "",
                                "PLAY [Apply role aodh] *********************************************************",
                                "",
                                "TASK [aodh : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:44:08 +0000 (0:00:00.450)       0:00:01.046 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/aodh/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Creating services] **************************",
                                "Thursday 26 March 2026  03:44:08 +0000 (0:00:00.622)       0:00:01.669 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=aodh (alarming))\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Creating endpoints] *************************",
                                "Thursday 26 March 2026  03:44:12 +0000 (0:00:03.760)       0:00:05.430 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> https://api-int.testbed.osism.xyz:8042 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> https://api.testbed.osism.xyz:8042 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Creating projects] **************************",
                                "Thursday 26 March 2026  03:44:19 +0000 (0:00:06.645)       0:00:12.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Creating users] *****************************",
                                "Thursday 26 March 2026  03:44:22 +0000 (0:00:03.452)       0:00:15.529 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Creating roles] *****************************",
                                "Thursday 26 March 2026  03:44:26 +0000 (0:00:03.925)       0:00:19.454 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : aodh | Granting user roles] ************************",
                                "Thursday 26 March 2026  03:44:29 +0000 (0:00:03.429)       0:00:22.884 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=aodh -> service -> admin)\u001b[0m",
                                "",
                                "TASK [aodh : Ensuring config directories exist] ********************************",
                                "Thursday 26 March 2026  03:44:33 +0000 (0:00:04.000)       0:00:26.885 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [aodh : Check if policies shall be overwritten] ***************************",
                                "Thursday 26 March 2026  03:44:36 +0000 (0:00:02.130)       0:00:29.016 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [aodh : Set aodh policy file] *********************************************",
                                "Thursday 26 March 2026  03:44:36 +0000 (0:00:00.132)       0:00:29.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [aodh : Copying over existing policy file] ********************************",
                                "Thursday 26 March 2026  03:44:36 +0000 (0:00:00.564)       0:00:29.712 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [aodh : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:44:37 +0000 (0:00:00.677)       0:00:30.390 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/aodh/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : aodh | Copying over extra CA certificates] ***********",
                                "Thursday 26 March 2026  03:44:38 +0000 (0:00:00.801)       0:00:31.192 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : aodh | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:44:42 +0000 (0:00:04.462)       0:00:35.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : aodh | Copying over backend internal TLS key] ********",
                                "Thursday 26 March 2026  03:44:43 +0000 (0:00:00.684)       0:00:36.338 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [aodh : Copying over config.json files for services] **********************",
                                "Thursday 26 March 2026  03:44:44 +0000 (0:00:01.107)       0:00:37.445 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [aodh : Copying over aodh.conf] *******************************************",
                                "Thursday 26 March 2026  03:44:48 +0000 (0:00:04.195)       0:00:41.641 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [aodh : Copying over wsgi-aodh files for services] ************************",
                                "Thursday 26 March 2026  03:44:57 +0000 (0:00:08.707)       0:00:50.348 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [aodh : Check aodh containers] ********************************************",
                                "Thursday 26 March 2026  03:44:59 +0000 (0:00:01.790)       0:00:52.138 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [aodh : include_tasks] ****************************************************",
                                "Thursday 26 March 2026  03:45:02 +0000 (0:00:03.375)       0:00:55.514 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [aodh : Creating aodh database] *******************************************",
                                "Thursday 26 March 2026  03:45:02 +0000 (0:00:00.335)       0:00:55.849 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [aodh : Creating aodh database user and setting permissions] **************",
                                "Thursday 26 March 2026  03:45:05 +0000 (0:00:02.153)       0:00:58.003 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [aodh : Running aodh bootstrap container] *********************************",
                                "Thursday 26 March 2026  03:45:07 +0000 (0:00:02.269)       0:01:00.273 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [aodh : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:45:20 +0000 (0:00:13.524)       0:01:13.797 ******** ",
                                "",
                                "TASK [aodh : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:45:20 +0000 (0:00:00.087)       0:01:13.884 ******** ",
                                "",
                                "TASK [aodh : Flush handlers] ***************************************************",
                                "Thursday 26 March 2026  03:45:20 +0000 (0:00:00.072)       0:01:13.957 ******** ",
                                "",
                                "RUNNING HANDLER [aodh : Restart aodh-api container] ****************************",
                                "Thursday 26 March 2026  03:45:21 +0000 (0:00:00.289)       0:01:14.246 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [aodh : Restart aodh-evaluator container] **********************",
                                "Thursday 26 March 2026  03:45:31 +0000 (0:00:10.640)       0:01:24.887 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [aodh : Restart aodh-listener container] ***********************",
                                "Thursday 26 March 2026  03:45:40 +0000 (0:00:08.230)       0:01:33.118 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [aodh : Restart aodh-notifier container] ***********************",
                                "Thursday 26 March 2026  03:45:50 +0000 (0:00:10.712)       0:01:43.831 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=17  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:46:01 +0000 (0:00:10.559)       0:01:54.390 ******** ",
                                "=============================================================================== ",
                                "aodh : Running aodh bootstrap container -------------------------------- 13.52s",
                                "aodh : Restart aodh-listener container --------------------------------- 10.71s",
                                "aodh : Restart aodh-api container -------------------------------------- 10.64s",
                                "aodh : Restart aodh-notifier container --------------------------------- 10.56s",
                                "aodh : Copying over aodh.conf ------------------------------------------- 8.71s",
                                "aodh : Restart aodh-evaluator container --------------------------------- 8.23s",
                                "service-ks-register : aodh | Creating endpoints ------------------------- 6.65s",
                                "service-cert-copy : aodh | Copying over extra CA certificates ----------- 4.46s",
                                "aodh : Copying over config.json files for services ---------------------- 4.20s",
                                "service-ks-register : aodh | Granting user roles ------------------------ 4.00s",
                                "service-ks-register : aodh | Creating users ----------------------------- 3.93s",
                                "service-ks-register : aodh | Creating services -------------------------- 3.76s",
                                "service-ks-register : aodh | Creating projects -------------------------- 3.45s",
                                "service-ks-register : aodh | Creating roles ----------------------------- 3.43s",
                                "aodh : Check aodh containers -------------------------------------------- 3.38s",
                                "aodh : Creating aodh database user and setting permissions -------------- 2.27s",
                                "aodh : Creating aodh database ------------------------------------------- 2.15s",
                                "aodh : Ensuring config directories exist -------------------------------- 2.13s",
                                "aodh : Copying over wsgi-aodh files for services ------------------------ 1.79s",
                                "service-cert-copy : aodh | Copying over backend internal TLS key -------- 1.11s",
                                "\u001b[32m2026-03-26 03:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac547b8e-a406-49a2-abee-9b9679420316 (kolla-ceph-rgw) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ac547b8e-a406-49a2-abee-9b9679420316 (kolla-ceph-rgw) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:46:08 +0000 (0:00:00.316)       0:00:00.316 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:46:09 +0000 (0:00:00.955)       0:00:01.272 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-rgw : include_tasks] ************************************************",
                                "Thursday 26 March 2026  03:46:10 +0000 (0:00:00.822)       0:00:02.094 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating services] **********************",
                                "Thursday 26 March 2026  03:46:12 +0000 (0:00:01.656)       0:00:03.751 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating endpoints] *********************",
                                "Thursday 26 March 2026  03:46:16 +0000 (0:00:03.922)       0:00:07.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating projects] **********************",
                                "Thursday 26 March 2026  03:46:22 +0000 (0:00:06.506)       0:00:14.180 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating users] *************************",
                                "Thursday 26 March 2026  03:46:25 +0000 (0:00:03.325)       0:00:17.505 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating roles] *************************",
                                "Thursday 26 March 2026  03:46:29 +0000 (0:00:04.015)       0:00:21.521 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Granting user roles] ********************",
                                "Thursday 26 March 2026  03:46:36 +0000 (0:00:06.487)       0:00:28.008 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:46:41 +0000 (0:00:05.109)       0:00:33.118 ******** ",
                                "=============================================================================== ",
                                "service-ks-register : ceph-rgw | Creating endpoints --------------------- 6.51s",
                                "service-ks-register : ceph-rgw | Creating roles ------------------------- 6.49s",
                                "service-ks-register : ceph-rgw | Granting user roles -------------------- 5.11s",
                                "service-ks-register : ceph-rgw | Creating users ------------------------- 4.02s",
                                "service-ks-register : ceph-rgw | Creating services ---------------------- 3.92s",
                                "service-ks-register : ceph-rgw | Creating projects ---------------------- 3.33s",
                                "ceph-rgw : include_tasks ------------------------------------------------ 1.66s",
                                "Group hosts based on Kolla action --------------------------------------- 0.96s",
                                "Group hosts based on enabled services ----------------------------------- 0.82s",
                                "\u001b[32m2026-03-26 03:46:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cab4e368-13de-4ef3-b76d-5a075b924f38 (gnocchi) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:46:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cab4e368-13de-4ef3-b76d-5a075b924f38 (gnocchi) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:46:49 +0000 (0:00:00.288)       0:00:00.288 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:46:49 +0000 (0:00:00.336)       0:00:00.625 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_gnocchi_False)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_gnocchi_True\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_gnocchi_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_gnocchi_False)\u001b[0m",
                                "",
                                "PLAY [Apply role gnocchi] ******************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:46:49 +0000 (0:00:00.415)       0:00:01.040 ******** ",
                                "=============================================================================== ",
                                "Group hosts based on enabled services ----------------------------------- 0.42s",
                                "Group hosts based on Kolla action --------------------------------------- 0.34s",
                                "\u001b[32m2026-03-26 03:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4dad45d8-ad3c-4a11-bb8f-ae77d3a40d56 (manila) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4dad45d8-ad3c-4a11-bb8f-ae77d3a40d56 (manila) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:46:57 +0000 (0:00:00.291)       0:00:00.291 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:46:57 +0000 (0:00:00.320)       0:00:00.611 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_manila_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_manila_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_manila_True)\u001b[0m",
                                "",
                                "PLAY [Apply role manila] *******************************************************",
                                "",
                                "TASK [manila : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:46:57 +0000 (0:00:00.466)       0:00:01.078 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/manila/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [manila : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:46:58 +0000 (0:00:00.575)       0:00:01.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Creating services] ************************",
                                "Thursday 26 March 2026  03:46:59 +0000 (0:00:00.494)       0:00:02.148 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila (share))\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 (sharev2))\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Creating endpoints] ***********************",
                                "Thursday 26 March 2026  03:47:05 +0000 (0:00:06.592)       0:00:08.741 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> https://api-int.testbed.osism.xyz:8786/v1/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> https://api.testbed.osism.xyz:8786/v1/%(tenant_id)s -> public)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 -> https://api-int.testbed.osism.xyz:8786/v2 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manilav2 -> https://api.testbed.osism.xyz:8786/v2 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Creating projects] ************************",
                                "Thursday 26 March 2026  03:47:18 +0000 (0:00:13.260)       0:00:22.002 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Creating users] ***************************",
                                "Thursday 26 March 2026  03:47:22 +0000 (0:00:03.416)       0:00:25.418 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Creating roles] ***************************",
                                "Thursday 26 March 2026  03:47:26 +0000 (0:00:03.981)       0:00:29.400 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : manila | Granting user roles] **********************",
                                "Thursday 26 March 2026  03:47:29 +0000 (0:00:03.226)       0:00:32.626 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila -> service -> admin)\u001b[0m",
                                "",
                                "TASK [manila : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  03:47:33 +0000 (0:00:04.011)       0:00:36.638 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [manila : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:47:35 +0000 (0:00:02.411)       0:00:39.049 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/manila/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [manila : Ensuring manila service ceph config subdir exists] **************",
                                "Thursday 26 March 2026  03:47:36 +0000 (0:00:00.633)       0:00:39.683 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [manila : Copy over multiple ceph configs for Manila] *********************",
                                "Thursday 26 March 2026  03:47:37 +0000 (0:00:01.053)       0:00:40.736 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "",
                                "TASK [manila : Copy over ceph Manila keyrings] *********************************",
                                "Thursday 26 March 2026  03:47:39 +0000 (0:00:01.877)       0:00:42.613 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'cephfsnative1', 'share_name': 'CEPHFS1', 'driver': 'cephfsnative', 'cluster': 'ceph', 'enabled': True, 'protocols': ['CEPHFS']})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cephfsnfs1', 'share_name': 'CEPHFSNFS1', 'driver': 'cephfsnfs', 'cluster': 'ceph', 'enabled': False, 'protocols': ['NFS', 'CIFS']}) \u001b[0m",
                                "",
                                "TASK [manila : Ensuring config directory has correct owner and permission] *****",
                                "Thursday 26 March 2026  03:47:40 +0000 (0:00:01.295)       0:00:43.909 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=manila-share)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=manila-share)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=manila-share)\u001b[0m",
                                "",
                                "TASK [manila : Check if policies shall be overwritten] *************************",
                                "Thursday 26 March 2026  03:47:41 +0000 (0:00:00.846)       0:00:44.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [manila : Set manila policy file] *****************************************",
                                "Thursday 26 March 2026  03:47:41 +0000 (0:00:00.155)       0:00:44.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [manila : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:47:42 +0000 (0:00:00.684)       0:00:45.595 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/manila/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : manila | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  03:47:43 +0000 (0:00:00.715)       0:00:46.310 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : manila | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:47:47 +0000 (0:00:04.172)       0:00:50.483 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : manila | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  03:47:48 +0000 (0:00:00.901)       0:00:51.384 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [manila : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  03:47:49 +0000 (0:00:00.911)       0:00:52.296 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [manila : Copying over manila.conf] ***************************************",
                                "Thursday 26 March 2026  03:47:53 +0000 (0:00:04.662)       0:00:56.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [manila : Copying over manila-share.conf] *********************************",
                                "Thursday 26 March 2026  03:48:00 +0000 (0:00:07.104)       0:01:04.063 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=manila-share)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=manila-share)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=manila-share)\u001b[0m",
                                "",
                                "TASK [manila : Copying over existing policy file] ******************************",
                                "Thursday 26 March 2026  03:48:04 +0000 (0:00:03.914)       0:01:07.977 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [manila : Check manila containers] ****************************************",
                                "Thursday 26 March 2026  03:48:05 +0000 (0:00:00.767)       0:01:08.745 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.1.1.20251130', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [manila : Creating Manila database] ***************************************",
                                "Thursday 26 March 2026  03:48:09 +0000 (0:00:03.570)       0:01:12.315 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [manila : Creating Manila database user and setting permissions] **********",
                                "Thursday 26 March 2026  03:48:11 +0000 (0:00:02.477)       0:01:14.793 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [manila : Running Manila bootstrap container] *****************************",
                                "Thursday 26 March 2026  03:48:14 +0000 (0:00:02.303)       0:01:17.096 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [manila : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:48:51 +0000 (0:00:37.486)       0:01:54.583 ******** ",
                                "",
                                "TASK [manila : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:48:51 +0000 (0:00:00.076)       0:01:54.660 ******** ",
                                "",
                                "TASK [manila : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  03:48:51 +0000 (0:00:00.074)       0:01:54.734 ******** ",
                                "",
                                "RUNNING HANDLER [manila : Restart manila-api container] ************************",
                                "Thursday 26 March 2026  03:48:51 +0000 (0:00:00.077)       0:01:54.811 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [manila : Restart manila-data container] ***********************",
                                "Thursday 26 March 2026  03:49:07 +0000 (0:00:15.489)       0:02:10.300 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [manila : Restart manila-scheduler container] ******************",
                                "Thursday 26 March 2026  03:49:17 +0000 (0:00:10.691)       0:02:20.992 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [manila : Restart manila-share container] **********************",
                                "Thursday 26 March 2026  03:49:28 +0000 (0:00:10.408)       0:02:31.400 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=28  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:49:47 +0000 (0:00:19.047)       0:02:50.448 ******** ",
                                "=============================================================================== ",
                                "manila : Running Manila bootstrap container ---------------------------- 37.49s",
                                "manila : Restart manila-share container -------------------------------- 19.05s",
                                "manila : Restart manila-api container ---------------------------------- 15.49s",
                                "service-ks-register : manila | Creating endpoints ---------------------- 13.26s",
                                "manila : Restart manila-data container --------------------------------- 10.69s",
                                "manila : Restart manila-scheduler container ---------------------------- 10.41s",
                                "manila : Copying over manila.conf --------------------------------------- 7.10s",
                                "service-ks-register : manila | Creating services ------------------------ 6.59s",
                                "manila : Copying over config.json files for services -------------------- 4.66s",
                                "service-cert-copy : manila | Copying over extra CA certificates --------- 4.17s",
                                "service-ks-register : manila | Granting user roles ---------------------- 4.01s",
                                "service-ks-register : manila | Creating users --------------------------- 3.98s",
                                "manila : Copying over manila-share.conf --------------------------------- 3.91s",
                                "manila : Check manila containers ---------------------------------------- 3.57s",
                                "service-ks-register : manila | Creating projects ------------------------ 3.42s",
                                "service-ks-register : manila | Creating roles --------------------------- 3.23s",
                                "manila : Creating Manila database --------------------------------------- 2.48s",
                                "manila : Ensuring config directories exist ------------------------------ 2.41s",
                                "manila : Creating Manila database user and setting permissions ---------- 2.30s",
                                "manila : Copy over multiple ceph configs for Manila --------------------- 1.88s",
                                "+ sh -c /opt/configuration/scripts/deploy/400-monitoring.sh",
                                "\u001b[32m2026-03-26 03:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0cae00d9-e756-4c7c-ac2c-a238d5320de9 (netdata) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0cae00d9-e756-4c7c-ac2c-a238d5320de9 (netdata) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:50:05 +0000 (0:00:00.303)       0:00:00.303 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m",
                                "",
                                "PLAY [Apply role netdata] ******************************************************",
                                "",
                                "TASK [osism.services.netdata : Include distribution specific install tasks] ****",
                                "Thursday 26 March 2026  03:50:06 +0000 (0:00:01.086)       0:00:01.389 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Remove old architecture-dependent repository] ***",
                                "Thursday 26 March 2026  03:50:08 +0000 (0:00:01.594)       0:00:02.984 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install apt-transport-https package] ************",
                                "Thursday 26 March 2026  03:50:10 +0000 (0:00:02.156)       0:00:05.141 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository gpg key] *************************",
                                "Thursday 26 March 2026  03:50:12 +0000 (0:00:02.351)       0:00:07.492 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository] *********************************",
                                "Thursday 26 March 2026  03:50:14 +0000 (0:00:01.655)       0:00:09.147 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install package netdata] ************************",
                                "Thursday 26 March 2026  03:50:30 +0000 (0:00:16.080)       0:00:25.228 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include config tasks] ***************************",
                                "Thursday 26 March 2026  03:51:12 +0000 (0:00:42.183)       0:01:07.411 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Copy configuration files] ***********************",
                                "Thursday 26 March 2026  03:51:14 +0000 (0:00:01.653)       0:01:09.064 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***",
                                "Thursday 26 March 2026  03:51:17 +0000 (0:00:03.435)       0:01:12.500 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Opt out from anonymous statistics] **************",
                                "Thursday 26 March 2026  03:51:19 +0000 (0:00:01.312)       0:01:13.812 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add netdata user to docker group] ***************",
                                "Thursday 26 March 2026  03:51:20 +0000 (0:00:01.334)       0:01:15.147 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Manage service netdata] *************************",
                                "Thursday 26 March 2026  03:51:21 +0000 (0:00:01.269)       0:01:16.417 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include host type specific tasks] ***************",
                                "Thursday 26 March 2026  03:51:23 +0000 (0:00:01.788)       0:01:18.205 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********",
                                "Thursday 26 March 2026  03:51:25 +0000 (0:00:01.523)       0:01:19.729 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.netdata : Restart service netdata] *************",
                                "Thursday 26 March 2026  03:51:28 +0000 (0:00:03.358)       0:01:23.087 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:51:39 +0000 (0:00:11.180)       0:01:34.268 ******** ",
                                "=============================================================================== ",
                                "osism.services.netdata : Install package netdata ----------------------- 42.18s",
                                "osism.services.netdata : Add repository -------------------------------- 16.08s",
                                "osism.services.netdata : Restart service netdata ----------------------- 11.18s",
                                "osism.services.netdata : Copy configuration files ----------------------- 3.44s",
                                "osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 3.36s",
                                "osism.services.netdata : Install apt-transport-https package ------------ 2.35s",
                                "osism.services.netdata : Remove old architecture-dependent repository --- 2.16s",
                                "osism.services.netdata : Manage service netdata ------------------------- 1.79s",
                                "osism.services.netdata : Add repository gpg key ------------------------- 1.66s",
                                "osism.services.netdata : Include config tasks --------------------------- 1.65s",
                                "osism.services.netdata : Include distribution specific install tasks ---- 1.59s",
                                "osism.services.netdata : Include host type specific tasks --------------- 1.52s",
                                "osism.services.netdata : Opt out from anonymous statistics -------------- 1.33s",
                                "osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.31s",
                                "osism.services.netdata : Add netdata user to docker group --------------- 1.27s",
                                "Group hosts based on enabled services ----------------------------------- 1.09s",
                                "\u001b[32m2026-03-26 03:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 841b3c3f-80c9-4ffe-85c0-5ab312ce4bb3 (prometheus) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 841b3c3f-80c9-4ffe-85c0-5ab312ce4bb3 (prometheus) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:51:50 +0000 (0:00:00.311)       0:00:00.311 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:51:52 +0000 (0:00:01.016)       0:00:01.328 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m",
                                "",
                                "PLAY [Apply role prometheus] ***************************************************",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Thursday 26 March 2026  03:51:52 +0000 (0:00:00.954)       0:00:02.282 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [prometheus : Ensuring config directories exist] **************************",
                                "Thursday 26 March 2026  03:51:54 +0000 (0:00:01.436)       0:00:03.718 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Thursday 26 March 2026  03:51:57 +0000 (0:00:03.021)       0:00:06.740 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****",
                                "Thursday 26 March 2026  03:51:59 +0000 (0:00:01.776)       0:00:08.516 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:52:04 +0000 (0:00:05.614)       0:00:14.130 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  03:52:06 +0000 (0:00:02.178)       0:00:16.309 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over config.json files] *****************************",
                                "Thursday 26 March 2026  03:52:09 +0000 (0:00:02.365)       0:00:18.675 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom prometheus alert rules files] *******************",
                                "Thursday 26 March 2026  03:52:16 +0000 (0:00:06.928)       0:00:25.603 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom prometheus alert rules files] ***********",
                                "Thursday 26 March 2026  03:52:17 +0000 (0:00:00.970)       0:00:26.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1100875, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8383117, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1100918, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8500881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1100857, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8349571, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1100901, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8452802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1100854, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8317897, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1100881, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8386195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1100896, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8417988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1100884, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8392801, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1100864, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83728, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100917, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100852, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1100931, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1100915, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8482802, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1100855, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83228, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1100853, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.83128, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1100893, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8408408, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1100887, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.84028, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1100928, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8532803, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus common config overrides] ********************",
                                "Thursday 26 March 2026  03:52:43 +0000 (0:00:25.818)       0:00:52.392 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus host config overrides] **********************",
                                "Thursday 26 March 2026  03:52:43 +0000 (0:00:00.819)       0:00:53.211 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus config file] ************************",
                                "Thursday 26 March 2026  03:52:45 +0000 (0:00:01.919)       0:00:55.131 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus web config file] ********************",
                                "Thursday 26 March 2026  03:53:03 +0000 (0:00:17.839)       0:01:12.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus alertmanager config file] ***********",
                                "Thursday 26 March 2026  03:53:06 +0000 (0:00:02.942)       0:01:15.913 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom Alertmanager alert notification templates] ******",
                                "Thursday 26 March 2026  03:53:08 +0000 (0:00:02.029)       0:01:17.943 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom Alertmanager alert notification templates] ***",
                                "Thursday 26 March 2026  03:53:09 +0000 (0:00:00.839)       0:01:18.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************",
                                "Thursday 26 March 2026  03:53:10 +0000 (0:00:00.868)       0:01:19.650 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying cloud config file for openstack exporter] ***********",
                                "Thursday 26 March 2026  03:53:12 +0000 (0:00:02.389)       0:01:22.040 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying config file for blackbox exporter] ******************",
                                "Thursday 26 March 2026  03:53:14 +0000 (0:00:01.614)       0:01:23.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Find extra prometheus server config files] ******************",
                                "Thursday 26 March 2026  03:53:16 +0000 (0:00:01.677)       0:01:25.332 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m",
                                "\u001b[1;35mnot a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Create subdirectories for extra config files] ***************",
                                "Thursday 26 March 2026  03:53:17 +0000 (0:00:01.186)       0:01:26.518 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Template extra prometheus server config files] **************",
                                "Thursday 26 March 2026  03:53:18 +0000 (0:00:01.015)       0:01:27.534 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Check prometheus containers] ********************************",
                                "Thursday 26 March 2026  03:53:19 +0000 (0:00:00.943)       0:01:28.478 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Creating prometheus database user and setting permissions] ***",
                                "Thursday 26 March 2026  03:53:23 +0000 (0:00:04.294)       0:01:32.772 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:24 +0000 (0:00:01.479)       0:01:34.252 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.082)       0:01:34.335 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.077)       0:01:34.412 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.074)       0:01:34.486 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.080)       0:01:34.567 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.070)       0:01:34.637 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.082)       0:01:34.719 ******** ",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-server container] *************",
                                "Thursday 26 March 2026  03:53:25 +0000 (0:00:00.097)       0:01:34.817 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******",
                                "Thursday 26 March 2026  03:53:47 +0000 (0:00:21.534)       0:01:56.352 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****",
                                "Thursday 26 March 2026  03:54:00 +0000 (0:00:13.813)       0:02:10.165 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***",
                                "Thursday 26 March 2026  03:54:10 +0000 (0:00:10.085)       0:02:20.251 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********",
                                "Thursday 26 March 2026  03:54:21 +0000 (0:00:10.498)       0:02:30.749 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******",
                                "Thursday 26 March 2026  03:54:36 +0000 (0:00:14.650)       0:02:45.400 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***",
                                "Thursday 26 March 2026  03:54:45 +0000 (0:00:08.962)       0:02:54.362 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***",
                                "Thursday 26 March 2026  03:54:50 +0000 (0:00:05.845)       0:03:00.208 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***",
                                "Thursday 26 March 2026  03:54:56 +0000 (0:00:05.838)       0:03:06.047 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:55:07 +0000 (0:00:10.524)       0:03:16.572 ******** ",
                                "=============================================================================== ",
                                "prometheus : Copying over custom prometheus alert rules files ---------- 25.82s",
                                "prometheus : Restart prometheus-server container ----------------------- 21.53s",
                                "prometheus : Copying over prometheus config file ----------------------- 17.84s",
                                "prometheus : Restart prometheus-cadvisor container --------------------- 14.65s",
                                "prometheus : Restart prometheus-node-exporter container ---------------- 13.81s",
                                "prometheus : Restart prometheus-libvirt-exporter container ------------- 10.52s",
                                "prometheus : Restart prometheus-memcached-exporter container ----------- 10.50s",
                                "prometheus : Restart prometheus-mysqld-exporter container -------------- 10.09s",
                                "prometheus : Restart prometheus-alertmanager container ------------------ 8.96s",
                                "prometheus : Copying over config.json files ----------------------------- 6.93s",
                                "prometheus : Restart prometheus-elasticsearch-exporter container -------- 5.85s",
                                "prometheus : Restart prometheus-blackbox-exporter container ------------- 5.84s",
                                "service-cert-copy : prometheus | Copying over extra CA certificates ----- 5.61s",
                                "prometheus : Check prometheus containers -------------------------------- 4.29s",
                                "prometheus : Ensuring config directories exist -------------------------- 3.02s",
                                "prometheus : Copying over prometheus web config file -------------------- 2.94s",
                                "prometheus : Copying over my.cnf for mysqld_exporter -------------------- 2.39s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS key --- 2.37s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.18s",
                                "prometheus : Copying over prometheus alertmanager config file ----------- 2.03s",
                                "\u001b[32m2026-03-26 03:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 00c4bf35-4c60-4b40-be13-05f79dea06c0 (grafana) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 00c4bf35-4c60-4b40-be13-05f79dea06c0 (grafana) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:55:14 +0000 (0:00:00.255)       0:00:00.255 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:55:14 +0000 (0:00:00.375)       0:00:00.630 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m",
                                "",
                                "PLAY [Apply role grafana] ******************************************************",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:55:14 +0000 (0:00:00.484)       0:00:01.114 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [grafana : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  03:55:15 +0000 (0:00:00.605)       0:00:01.720 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Check if extra configuration file exists] **********************",
                                "Thursday 26 March 2026  03:55:16 +0000 (0:00:00.923)       0:00:02.643 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Thursday 26 March 2026  03:55:17 +0000 (0:00:00.910)       0:00:03.554 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  03:55:18 +0000 (0:00:00.642)       0:00:04.196 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:55:19 +0000 (0:00:01.379)       0:00:05.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  03:55:20 +0000 (0:00:00.638)       0:00:06.214 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over config.json files] ********************************",
                                "Thursday 26 March 2026  03:55:20 +0000 (0:00:00.666)       0:00:06.881 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over grafana.ini] **************************************",
                                "Thursday 26 March 2026  03:55:22 +0000 (0:00:01.369)       0:00:08.251 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over extra configuration file] *************************",
                                "Thursday 26 March 2026  03:55:23 +0000 (0:00:01.769)       0:00:10.021 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Configuring Prometheus as data source for Grafana] *************",
                                "Thursday 26 March 2026  03:55:24 +0000 (0:00:00.390)       0:00:10.411 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "",
                                "TASK [grafana : Configuring dashboards provisioning] ***************************",
                                "Thursday 26 March 2026  03:55:25 +0000 (0:00:01.360)       0:00:11.771 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "",
                                "TASK [grafana : Find custom grafana dashboards] ********************************",
                                "Thursday 26 March 2026  03:55:27 +0000 (0:00:01.892)       0:00:13.664 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : Find templated grafana dashboards] *****************************",
                                "Thursday 26 March 2026  03:55:28 +0000 (0:00:00.809)       0:00:14.474 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Prune templated Grafana dashboards] ****************************",
                                "Thursday 26 March 2026  03:55:29 +0000 (0:00:00.739)       0:00:15.213 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over custom dashboards] ********************************",
                                "Thursday 26 March 2026  03:55:29 +0000 (0:00:00.378)       0:00:15.593 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1100632, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.732278, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1100710, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1100655, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7379594, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1100712, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.751896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1100679, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7423134, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1100701, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7472785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1100630, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.731227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1100641, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7362783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1100659, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7392783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1100691, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744052, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1100709, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7492785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1100651, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7372782, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1100698, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7450945, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1100686, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7433124, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1100676, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.741759, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1100672, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7411568, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1100694, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.744788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1100665, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7403498, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1100707, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7482784, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1100847, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.82828, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1100750, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.783826, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1100729, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.756929, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1100796, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7882793, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1100722, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7527258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1100838, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8142796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1100797, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8092794, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1100839, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8152797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1100844, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8262799, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1100837, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8132796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1100790, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1100747, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.761357, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1100787, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.784279, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1100739, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7602787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1100793, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7862792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1100842, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8242798, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1100841, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8192797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1100724, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7542787, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1100726, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.7552786, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1100831, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8122797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1100840, 'dev': 145, 'nlink': 1, 'atime': 1764530892.0, 'mtime': 1764530892.0, 'ctime': 1774489902.8162796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "",
                                "TASK [grafana : Check grafana containers] **************************************",
                                "Thursday 26 March 2026  03:56:08 +0000 (0:00:38.713)       0:00:54.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.3.0.20251130', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database] *************************************",
                                "Thursday 26 March 2026  03:56:09 +0000 (0:00:01.098)       0:00:55.404 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database user and setting permissions] ********",
                                "Thursday 26 March 2026  03:56:11 +0000 (0:00:02.371)       0:00:57.776 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:56:14 +0000 (0:00:02.393)       0:01:00.169 ******** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:56:14 +0000 (0:00:00.094)       0:01:00.264 ******** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Thursday 26 March 2026  03:56:14 +0000 (0:00:00.106)       0:01:00.370 ******** ",
                                "",
                                "RUNNING HANDLER [grafana : Restart first grafana container] ********************",
                                "Thursday 26 March 2026  03:56:14 +0000 (0:00:00.085)       0:01:00.456 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********",
                                "Thursday 26 March 2026  03:56:16 +0000 (0:00:02.293)       0:01:02.749 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (9 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Restart remaining grafana containers] ***************",
                                "Thursday 26 March 2026  03:57:07 +0000 (0:00:51.148)       0:01:53.898 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Wait for grafana application ready] ****************************",
                                "Thursday 26 March 2026  03:57:41 +0000 (0:00:33.842)       0:02:27.740 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Remove old grafana docker volume] ******************************",
                                "Thursday 26 March 2026  03:57:44 +0000 (0:00:02.374)       0:02:30.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Enable grafana datasources] ************************************",
                                "Thursday 26 March 2026  03:57:44 +0000 (0:00:00.356)       0:02:30.471 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Disable Getting Started panel] *********************************",
                                "Thursday 26 March 2026  03:57:47 +0000 (0:00:02.724)       0:02:33.195 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:57:47 +0000 (0:00:00.328)       0:02:33.523 ******** ",
                                "=============================================================================== ",
                                "grafana : Waiting for grafana to start on first node ------------------- 51.15s",
                                "grafana : Copying over custom dashboards ------------------------------- 38.71s",
                                "grafana : Restart remaining grafana containers ------------------------- 33.84s",
                                "grafana : Enable grafana datasources ------------------------------------ 2.72s",
                                "grafana : Creating grafana database user and setting permissions -------- 2.39s",
                                "grafana : Wait for grafana application ready ---------------------------- 2.37s",
                                "grafana : Creating grafana database ------------------------------------- 2.37s",
                                "grafana : Restart first grafana container ------------------------------- 2.29s",
                                "grafana : Configuring dashboards provisioning --------------------------- 1.89s",
                                "grafana : Copying over grafana.ini -------------------------------------- 1.77s",
                                "service-cert-copy : grafana | Copying over extra CA certificates -------- 1.38s",
                                "grafana : Copying over config.json files -------------------------------- 1.37s",
                                "grafana : Configuring Prometheus as data source for Grafana ------------- 1.36s",
                                "grafana : Check grafana containers -------------------------------------- 1.10s",
                                "grafana : Ensuring config directories exist ----------------------------- 0.92s",
                                "grafana : Check if extra configuration file exists ---------------------- 0.91s",
                                "grafana : Find custom grafana dashboards -------------------------------- 0.81s",
                                "grafana : Find templated grafana dashboards ----------------------------- 0.74s",
                                "service-cert-copy : grafana | Copying over backend internal TLS key ----- 0.67s",
                                "grafana : include_tasks ------------------------------------------------- 0.64s",
                                "+ sh -c /opt/configuration/scripts/deploy/510-clusterapi.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "++ semver 9.5.0 8.0.0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ osism apply clusterapi",
                                "\u001b[32m2026-03-26 03:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8d0762b2-3f92-45d4-a9d6-72554dc03cb3 (clusterapi) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8d0762b2-3f92-45d4-a9d6-72554dc03cb3 (clusterapi) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply cert_manager role] *************************************************",
                                "",
                                "TASK [Include cert_manager role] ***********************************************",
                                "Thursday 26 March 2026  03:57:55 +0000 (0:00:00.208)       0:00:00.208 ******** ",
                                "\u001b[0;36mincluded: cert_manager for testbed-manager\u001b[0m",
                                "",
                                "TASK [cert_manager : Deploy cert-manager crds] *********************************",
                                "Thursday 26 March 2026  03:57:56 +0000 (0:00:00.262)       0:00:00.470 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [cert_manager : Deploy cert-manager] **************************************",
                                "Thursday 26 March 2026  03:58:01 +0000 (0:00:05.815)       0:00:06.285 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Initialize or upgrade the CAPI management cluster] ***********************",
                                "",
                                "TASK [Get capi-system namespace phase] *****************************************",
                                "Thursday 26 March 2026  03:58:26 +0000 (0:00:24.656)       0:00:30.942 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set capi-system-phase fact] **********************************************",
                                "Thursday 26 March 2026  03:58:27 +0000 (0:00:01.152)       0:00:32.094 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Initialize the CAPI management cluster] **********************************",
                                "Thursday 26 March 2026  03:58:27 +0000 (0:00:00.170)       0:00:32.265 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Upgrade the CAPI management cluster] *************************************",
                                "Thursday 26 March 2026  03:58:44 +0000 (0:00:16.373)       0:00:48.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Install openstack-resource-controller] ***********************************",
                                "Thursday 26 March 2026  03:58:44 +0000 (0:00:00.169)       0:00:48.808 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  03:58:46 +0000 (0:00:02.321)       0:00:51.130 ******** ",
                                "=============================================================================== ",
                                "cert_manager : Deploy cert-manager ------------------------------------- 24.66s",
                                "Initialize the CAPI management cluster --------------------------------- 16.37s",
                                "cert_manager : Deploy cert-manager crds --------------------------------- 5.82s",
                                "Install openstack-resource-controller ----------------------------------- 2.32s",
                                "Get capi-system namespace phase ----------------------------------------- 1.15s",
                                "Include cert_manager role ----------------------------------------------- 0.26s",
                                "Set capi-system-phase fact ---------------------------------------------- 0.17s",
                                "Upgrade the CAPI management cluster ------------------------------------- 0.17s",
                                "+ osism apply magnum",
                                "\u001b[32m2026-03-26 03:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 74b26c9e-e173-4941-b638-fff6166fc4d3 (magnum) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 03:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 74b26c9e-e173-4941-b638-fff6166fc4d3 (magnum) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  03:58:54 +0000 (0:00:00.301)       0:00:00.301 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  03:58:54 +0000 (0:00:00.336)       0:00:00.638 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m",
                                "",
                                "PLAY [Apply role magnum] *******************************************************",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:58:55 +0000 (0:00:00.529)       0:00:01.167 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating services] ************************",
                                "Thursday 26 March 2026  03:58:55 +0000 (0:00:00.669)       0:00:01.837 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating endpoints] ***********************",
                                "Thursday 26 March 2026  03:58:59 +0000 (0:00:03.626)       0:00:05.464 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating projects] ************************",
                                "Thursday 26 March 2026  03:59:05 +0000 (0:00:06.472)       0:00:11.936 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating users] ***************************",
                                "Thursday 26 March 2026  03:59:09 +0000 (0:00:03.545)       0:00:15.482 ******** ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating roles] ***************************",
                                "Thursday 26 March 2026  03:59:13 +0000 (0:00:03.939)       0:00:19.422 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Granting user roles] **********************",
                                "Thursday 26 March 2026  03:59:16 +0000 (0:00:03.416)       0:00:22.838 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee domain] *********************************",
                                "Thursday 26 March 2026  03:59:21 +0000 (0:00:04.371)       0:00:27.210 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user] ***********************************",
                                "Thursday 26 March 2026  03:59:24 +0000 (0:00:03.622)       0:00:30.832 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user role] ******************************",
                                "Thursday 26 March 2026  03:59:28 +0000 (0:00:03.854)       0:00:34.687 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  03:59:32 +0000 (0:00:03.448)       0:00:38.135 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Check if policies shall be overwritten] *************************",
                                "Thursday 26 March 2026  03:59:33 +0000 (0:00:01.669)       0:00:39.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum policy file] *****************************************",
                                "Thursday 26 March 2026  03:59:33 +0000 (0:00:00.189)       0:00:39.995 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check if kubeconfig file is supplied] ***************************",
                                "Thursday 26 March 2026  03:59:34 +0000 (0:00:00.341)       0:00:40.336 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over kubeconfig file] ***********************************",
                                "Thursday 26 March 2026  03:59:35 +0000 (0:00:00.911)       0:00:41.248 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum kubeconfig file's path] ******************************",
                                "Thursday 26 March 2026  03:59:37 +0000 (0:00:02.571)       0:00:43.819 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:59:38 +0000 (0:00:00.548)       0:00:44.367 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  03:59:38 +0000 (0:00:00.582)       0:00:44.950 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  03:59:41 +0000 (0:00:02.655)       0:00:47.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  03:59:42 +0000 (0:00:01.094)       0:00:48.699 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  03:59:43 +0000 (0:00:00.934)       0:00:49.633 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over magnum.conf] ***************************************",
                                "Thursday 26 March 2026  03:59:46 +0000 (0:00:02.739)       0:00:52.373 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over existing policy file] ******************************",
                                "Thursday 26 March 2026  03:59:51 +0000 (0:00:05.740)       0:00:58.113 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check magnum containers] ****************************************",
                                "Thursday 26 March 2026  03:59:52 +0000 (0:00:00.797)       0:00:58.910 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Thursday 26 March 2026  03:59:54 +0000 (0:00:01.979)       0:01:00.889 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database] ***************************************",
                                "Thursday 26 March 2026  03:59:55 +0000 (0:00:00.754)       0:01:01.644 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database user and setting permissions] **********",
                                "Thursday 26 March 2026  03:59:57 +0000 (0:00:02.176)       0:01:03.820 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Running Magnum bootstrap container] *****************************",
                                "Thursday 26 March 2026  04:00:00 +0000 (0:00:02.444)       0:01:06.265 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:00:17 +0000 (0:00:17.263)       0:01:23.529 ******** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:00:17 +0000 (0:00:00.089)       0:01:23.618 ******** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:00:17 +0000 (0:00:00.076)       0:01:23.695 ******** ",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-api container] ************************",
                                "Thursday 26 March 2026  04:00:17 +0000 (0:00:00.075)       0:01:23.771 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-conductor container] ******************",
                                "Thursday 26 March 2026  04:00:38 +0000 (0:00:20.763)       0:01:44.534 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:00:50 +0000 (0:00:11.676)       0:01:56.211 ******** ",
                                "=============================================================================== ",
                                "magnum : Restart magnum-api container ---------------------------------- 20.76s",
                                "magnum : Running Magnum bootstrap container ---------------------------- 17.26s",
                                "magnum : Restart magnum-conductor container ---------------------------- 11.68s",
                                "service-ks-register : magnum | Creating endpoints ----------------------- 6.47s",
                                "magnum : Copying over magnum.conf --------------------------------------- 5.74s",
                                "service-ks-register : magnum | Granting user roles ---------------------- 4.37s",
                                "service-ks-register : magnum | Creating users --------------------------- 3.94s",
                                "magnum : Creating Magnum trustee user ----------------------------------- 3.85s",
                                "service-ks-register : magnum | Creating services ------------------------ 3.63s",
                                "magnum : Creating Magnum trustee domain --------------------------------- 3.62s",
                                "service-ks-register : magnum | Creating projects ------------------------ 3.55s",
                                "magnum : Creating Magnum trustee user role ------------------------------ 3.45s",
                                "service-ks-register : magnum | Creating roles --------------------------- 3.42s",
                                "magnum : Copying over config.json files for services -------------------- 2.74s",
                                "service-cert-copy : magnum | Copying over extra CA certificates --------- 2.66s",
                                "magnum : Copying over kubeconfig file ----------------------------------- 2.57s",
                                "magnum : Creating Magnum database user and setting permissions ---------- 2.44s",
                                "magnum : Creating Magnum database --------------------------------------- 2.18s",
                                "magnum : Check magnum containers ---------------------------------------- 1.98s",
                                "magnum : Ensuring config directories exist ------------------------------ 1.67s"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001c-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:00:51.462844Z",
                            "start": "2026-03-26T02:15:22.411215Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001c",
                        "name": "Deploy services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_nutshell | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:00:52.148554Z",
                            "start": "2026-03-26T04:00:51.637938Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001d",
                        "name": "Deploy in a nutshell"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "/opt/configuration/scripts/bootstrap.sh"
                            ],
                            "delta": "0:03:12.839139",
                            "end": "2026-03-26 04:04:05.320851",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 /opt/configuration/scripts/bootstrap.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001e-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 04:00:52.481712",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# BOOTSTRAP\n\n+ set -e\n+ echo\n+ echo '# BOOTSTRAP'\n+ echo\n+ sh -c /opt/configuration/scripts/bootstrap-services.sh\n+ set -e\n+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh\n\u001b[32m2026-03-26 04:00:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 31b0c09a-7ac3-467c-be7c-a759484498ca (flavor-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-26 04:00:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m\n\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m\n\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m\n\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m\n\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m\n\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m\n\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m\n\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m\n\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m\n\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m\n\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m\n\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m\n\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m\n\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m\n\u001b[32m2026-03-26 04:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m\n\u001b[32m2026-03-26 04:01:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 23a49ab9-b06a-451c-9053-ebcf8ce10483 (bootstrap-basic) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:01:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 23a49ab9-b06a-451c-9053-ebcf8ce10483 (bootstrap-basic) has been started and output is visible here.\u001b[0m\n\nPLAY [Bootstrap basic OpenStack services] **************************************\n\nTASK [Gathering Facts] *********************************************************\nThursday 26 March 2026  04:01:22 +0000 (0:00:00.092)       0:00:00.092 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Get volume type LUKS] ****************************************************\nThursday 26 March 2026  04:01:24 +0000 (0:00:02.112)       0:00:02.205 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Create volume type LUKS] *************************************************\nThursday 26 March 2026  04:01:32 +0000 (0:00:07.687)       0:00:09.892 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public network] ***************************************************\nThursday 26 March 2026  04:01:38 +0000 (0:00:06.342)       0:00:16.234 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Set public network to default] *******************************************\nThursday 26 March 2026  04:01:44 +0000 (0:00:05.625)       0:00:21.860 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public subnet] ****************************************************\nThursday 26 March 2026  04:01:50 +0000 (0:00:06.518)       0:00:28.378 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create default IPv4 subnet pool] *****************************************\nThursday 26 March 2026  04:01:54 +0000 (0:00:04.320)       0:00:32.699 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create manager role] *****************************************************\nThursday 26 March 2026  04:01:58 +0000 (0:00:03.967)       0:00:36.667 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:02:02 +0000 (0:00:03.540)       0:00:40.207 ******** \n=============================================================================== \nGet volume type LUKS ---------------------------------------------------- 7.69s\nSet public network to default ------------------------------------------- 6.52s\nCreate volume type LUKS ------------------------------------------------- 6.34s\nCreate public network --------------------------------------------------- 5.63s\nCreate public subnet ---------------------------------------------------- 4.32s\nCreate default IPv4 subnet pool ----------------------------------------- 3.97s\nCreate manager role ----------------------------------------------------- 3.54s\nGathering Facts --------------------------------------------------------- 2.11s\n\u001b[32m2026-03-26 04:02:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b74b2da2-3ef3-4167-9a51-4ffcf0b7b791 (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-26 04:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m\n\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-26 04:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-26 04:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m\n\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-26 04:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-26 04:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:02:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m\n\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m\n+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2.CHECKSUM\u001b[0m\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 95322c96815879973a320f11cf6c9ad6237f8791183c852e0c8319e08839b1ac\u001b[0m\n\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 025d7cac-e919-43a4-aaa9-75a2a23da910 (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m\n\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2: 200\u001b[0m\n\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m\n\u001b[32m2026-03-26 04:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-26 04:02:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-26' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m\n\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m\n\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m\n\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m\n\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m\n\u001b[32m2026-03-26 04:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m\n\u001b[32m2026-03-26 04:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m\n\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m\n\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m\n\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-26\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m\n\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m",
                            "stdout_lines": [
                                "",
                                "# BOOTSTRAP",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# BOOTSTRAP'",
                                "+ echo",
                                "+ sh -c /opt/configuration/scripts/bootstrap-services.sh",
                                "+ set -e",
                                "+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh",
                                "\u001b[32m2026-03-26 04:00:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 31b0c09a-7ac3-467c-be7c-a759484498ca (flavor-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 23a49ab9-b06a-451c-9053-ebcf8ce10483 (bootstrap-basic) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:01:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 23a49ab9-b06a-451c-9053-ebcf8ce10483 (bootstrap-basic) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Bootstrap basic OpenStack services] **************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Thursday 26 March 2026  04:01:22 +0000 (0:00:00.092)       0:00:00.092 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Get volume type LUKS] ****************************************************",
                                "Thursday 26 March 2026  04:01:24 +0000 (0:00:02.112)       0:00:02.205 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Create volume type LUKS] *************************************************",
                                "Thursday 26 March 2026  04:01:32 +0000 (0:00:07.687)       0:00:09.892 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public network] ***************************************************",
                                "Thursday 26 March 2026  04:01:38 +0000 (0:00:06.342)       0:00:16.234 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Set public network to default] *******************************************",
                                "Thursday 26 March 2026  04:01:44 +0000 (0:00:05.625)       0:00:21.860 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public subnet] ****************************************************",
                                "Thursday 26 March 2026  04:01:50 +0000 (0:00:06.518)       0:00:28.378 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create default IPv4 subnet pool] *****************************************",
                                "Thursday 26 March 2026  04:01:54 +0000 (0:00:04.320)       0:00:32.699 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create manager role] *****************************************************",
                                "Thursday 26 March 2026  04:01:58 +0000 (0:00:03.967)       0:00:36.667 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:02:02 +0000 (0:00:03.540)       0:00:40.207 ******** ",
                                "=============================================================================== ",
                                "Get volume type LUKS ---------------------------------------------------- 7.69s",
                                "Set public network to default ------------------------------------------- 6.52s",
                                "Create volume type LUKS ------------------------------------------------- 6.34s",
                                "Create public network --------------------------------------------------- 5.63s",
                                "Create public subnet ---------------------------------------------------- 4.32s",
                                "Create default IPv4 subnet pool ----------------------------------------- 3.97s",
                                "Create manager role ----------------------------------------------------- 3.54s",
                                "Gathering Facts --------------------------------------------------------- 2.11s",
                                "\u001b[32m2026-03-26 04:02:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b74b2da2-3ef3-4167-9a51-4ffcf0b7b791 (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2.CHECKSUM\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 95322c96815879973a320f11cf6c9ad6237f8791183c852e0c8319e08839b1ac\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 025d7cac-e919-43a4-aaa9-75a2a23da910 (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2: 200\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-26 04:02:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-26' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260326.qcow2\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-26\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-26'\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:04\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001e-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:04:05.375512Z",
                            "start": "2026-03-26T04:00:52.198681Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001e",
                        "name": "Bootstrap services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "/opt/configuration/scripts/check.sh"
                            ],
                            "delta": "0:10:52.001766",
                            "end": "2026-03-26 04:14:57.732050",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 /opt/configuration/scripts/check.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001f-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 04:04:05.730284",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# CHECK\n\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n+ echo\n+ echo '# CHECK'\n+ echo\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver 9.5.0 5.0.0\n\n## Containers @ testbed-manager\n\n+ [[ 1 -eq -1 ]]\n+ echo\n+ echo '## Containers @ testbed-manager'\n+ echo\n+ osism container testbed-manager ps\n\u001b[32m2026-03-26 04:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m\nCONTAINER ID   IMAGE                                                                            COMMAND                  CREATED             STATUS                 PORTS                                                                             NAMES\na6c0ff0cf266   registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_blackbox_exporter\n9950518a1d95   registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130        \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_alertmanager\n2d772ec2381d   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130            \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_cadvisor\n3db09ddd1147   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130        \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                                                                                            prometheus_node_exporter\n9171c7f54c01   registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130           \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                                                                                            prometheus_server\n42b99ccdba78   registry.osism.tech/osism/cephclient:18.2.7                                      \"/usr/bin/dumb-init \u2026\"   About an hour ago   Up 59 minutes                                                                                            cephclient\n14fa8b69c69f   registry.osism.tech/kolla/release/cron:3.0.20251130                              \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               cron\ne7d025959a4d   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               kolla_toolbox\n76b296741eb4   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               fluentd\n7e42ae5d1d9f   registry.osism.tech/osism/openstackclient:2024.2                                 \"/usr/bin/dumb-init \u2026\"   2 hours ago         Up 2 hours                                                                                               openstackclient\n9c516885849a   phpmyadmin/phpmyadmin:5.2                                                        \"/docker-entrypoint.\u2026\"   2 hours ago         Up 2 hours (healthy)   80/tcp                                                                            phpmyadmin\n218b44791923   registry.osism.tech/osism/homer:v25.10.1                                         \"/bin/sh /entrypoint\u2026\"   2 hours ago         Up 2 hours (healthy)   8080/tcp                                                                          homer\n26eeba89e7d3   registry.osism.tech/osism/cgit:1.2.3                                             \"httpd-foreground\"       2 hours ago         Up 2 hours             80/tcp                                                                            cgit\n01f240bfb490   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta                        \"entrypoint.sh -f /e\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:3128->3128/tcp                                                       squid\n46bc210ef4f1   registry.osism.tech/osism/inventory-reconciler:0.20251130.0                      \"/sbin/tini -- /entr\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-inventory_reconciler-1\ne5947914229b   registry.osism.tech/osism/ceph-ansible:0.20251130.0                              \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     ceph-ansible\n80f4f4ccb433   registry.osism.tech/osism/osism-ansible:0.20251130.0                             \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osism-ansible\n4cce80abddf5   registry.osism.tech/osism/osism-kubernetes:0.20251130.0                          \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osism-kubernetes\n383c6d4109e1   registry.osism.tech/osism/kolla-ansible:0.20251130.0                             \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     kolla-ansible\n73c5df6619ee   registry.osism.tech/osism/ara-server:1.7.3                                       \"sh -c '/wait && /ru\u2026\"   2 hours ago         Up 2 hours (healthy)   8000/tcp                                                                          manager-ara-server-1\na1385669c48e   registry.osism.tech/osism/osism-frontend:0.20251130.1                            \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours             192.168.16.5:3000->3000/tcp                                                       osism-frontend\naa27c19ce640   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-listener-1\n319e06da1161   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-flower-1\n0084ba6d7452   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-beat-1\n61ee0d48fd3c   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1\nc5005444125f   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- sleep\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osismclient\nfa875137e67a   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine                         \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours (healthy)   6379/tcp                                                                          manager-redis-1\n4bd574e40d14   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-openstack-1\nee921f5b5733   registry.osism.tech/dockerhub/library/mariadb:11.8.4                             \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours (healthy)   3306/tcp                                                                          manager-mariadb-1\n9ba1921448df   registry.osism.tech/dockerhub/library/traefik:v3.5.0                             \"/entrypoint.sh trae\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik\n\n## Images @ testbed-manager\n\n+ echo\n+ echo '## Images @ testbed-manager'\n+ echo\n+ osism container testbed-manager images\nREPOSITORY                                                       TAG               IMAGE ID       CREATED         SIZE\nregistry.osism.tech/osism/openstackclient                        2024.2            1806a5037062   24 hours ago    239MB\nregistry.osism.tech/dockerhub/library/redis                      7.4.7-alpine      e08bd8d5a677   8 weeks ago     41.4MB\nregistry.osism.tech/osism/homer                                  v25.10.1          ea34b371c716   3 months ago    11.5MB\nregistry.osism.tech/osism/kolla-ansible                          0.20251130.0      0f140ec71e5f   3 months ago    608MB\nregistry.osism.tech/kolla/release/kolla-toolbox                  19.7.1.20251130   314d22193a72   3 months ago    669MB\nregistry.osism.tech/kolla/release/cron                           3.0.20251130      e1e0428a330f   3 months ago    265MB\nregistry.osism.tech/kolla/release/fluentd                        5.0.8.20251130    fb3c98fc8cae   3 months ago    578MB\nregistry.osism.tech/kolla/release/prometheus-blackbox-exporter   0.25.0.20251130   7bbb4f6f4831   3 months ago    308MB\nregistry.osism.tech/kolla/release/prometheus-cadvisor            0.49.2.20251130   591cbce746c1   3 months ago    357MB\nregistry.osism.tech/kolla/release/prometheus-alertmanager        0.28.0.20251130   ba994ea4acda   3 months ago    404MB\nregistry.osism.tech/kolla/release/prometheus-v2-server           2.55.1.20251130   56b43d5c716a   3 months ago    839MB\nregistry.osism.tech/kolla/release/prometheus-node-exporter       1.8.2.20251130    c1ab1d07f7ef   3 months ago    305MB\nregistry.osism.tech/osism/inventory-reconciler                   0.20251130.0      1bfc1dadeee1   3 months ago    330MB\nregistry.osism.tech/osism/osism-ansible                          0.20251130.0      42988b2d229c   3 months ago    613MB\nregistry.osism.tech/osism/ceph-ansible                           0.20251130.0      a212d8ca4a50   3 months ago    560MB\nregistry.osism.tech/osism/osism-kubernetes                       0.20251130.0      9beff03cb77b   3 months ago    1.23GB\nregistry.osism.tech/osism/osism                                  0.20251130.1      95213af683ec   3 months ago    383MB\nregistry.osism.tech/osism/osism-frontend                         0.20251130.1      2cb6e7609620   3 months ago    238MB\nregistry.osism.tech/dockerhub/library/mariadb                    11.8.4            70745dd8f1d0   4 months ago    334MB\nphpmyadmin/phpmyadmin                                            5.2               e66b1f5a8c58   5 months ago    742MB\nregistry.osism.tech/osism/ara-server                             1.7.3             d1b687333f2f   7 months ago    275MB\nregistry.osism.tech/dockerhub/library/traefik                    v3.5.0            11cc59587f6a   8 months ago    226MB\nregistry.osism.tech/osism/cephclient                             18.2.7            ae977aa79826   10 months ago   453MB\nregistry.osism.tech/dockerhub/ubuntu/squid                       6.1-23.10_beta    34b6bbbcf74b   21 months ago   146MB\nregistry.osism.tech/osism/cgit                                   1.2.3             16e7285642b1   2 years ago     545MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver 9.5.0 5.0.0\n\n## Containers @ testbed-node-0\n\n+ [[ 1 -eq -1 ]]\n+ echo\n+ echo '## Containers @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 ps\nCONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES\n10e685d017e5   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor\n1e8d61047d06   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api\ndb6b2a95a9bc   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   7 minutes ago       Up 7 minutes                           grafana\n6542f78de3af   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter\n467c18e96c3a   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor\n137fab20c985   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_memcached_exporter\n937717e9414d   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter\n7eb22a81cd5e   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter\nf3fd6a1d9bdd   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share\n7f043dc613c0   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_scheduler\n7d2ea26d8a03   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data\nb1dd7de43631   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api\n122590345af5   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier\nc6af9a1009fc   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener\n3a51fd5dcb26   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator\n212c65b505ce   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api\n428298fcfb6e   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central\n6fff7b3529d3   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification\ne0fc6d6cd546   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   21 minutes ago      Up 21 minutes (healthy)                octavia_worker\n658463a5ab16   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping\n1015fd0d4be5   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager\n938dfb4b5809   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent\n7b61d18683f1   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api\n84d11367df1c   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker\nf53fd21463f1   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns\n6d3d7a997c3a   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer\n7afbf1ab3e1d   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central\nd021acb82cd1   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api\nbbf56f666257   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9\n02345889a066   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker\n6690561e229b   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_keystone_listener\nf3e2f038b0cd   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api\n7eacd3eb081d   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup\n33e9db06bbef   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_volume\n501c5dc0acdd   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler\n15d969d91992   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api\nb158cf01e40b   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api\nee404b650bd7   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console\n29a44929d5a6   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver\nf16c1f62e400   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon\n2415fb04f136   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy\n7886f78fe202   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor\n156eb6887714   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   44 minutes ago      Up 44 minutes (healthy)                nova_api\nab5d246d0721   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler\n4ad302f098f7   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server\n139e866a4b61   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   53 minutes ago      Up 53 minutes (healthy)                placement_api\n1c6ed65052b6   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 55 minutes (healthy)                keystone\nd393c5606814   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet\nf820741ba15d   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh\n198d679236a4   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-0\n48ac09de19f5   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-0\nc1b85917b265   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-0\n28734fa9b704   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd\n83ce56dde60c   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db\n5d67cc327d74   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db\nfdf645f7da16   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller\n4e815f33dd1c   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd\n5d315a0b0023   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db\n196668baf1fa   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq\n7d42071b582a   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb\nb05b43a972e2   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel\ne9731f9eb885   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis\n9b8a2aa0aa07   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached\nb1cb8702fe0a   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch_dashboards\n05b2989d837f   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch\nbb1665a8bd8c   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived\n2f3893ca35e0   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql\n9ccbd30640c7   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy\n97d9ef633eb8   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron\ncd160ef7c762   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox\ndc42d1f63132   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd\n\n## Images @ testbed-node-0\n\n+ echo\n+ echo '## Images @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 images\nREPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE\nregistry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB\nregistry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB\nregistry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB\nregistry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB\nregistry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB\nregistry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB\nregistry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB\nregistry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB\nregistry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB\nregistry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB\nregistry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB\nregistry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB\nregistry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB\nregistry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB\nregistry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB\nregistry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB\nregistry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB\nregistry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB\nregistry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB\nregistry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB\nregistry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB\nregistry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB\nregistry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB\nregistry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB\nregistry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB\nregistry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB\nregistry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB\nregistry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB\nregistry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB\nregistry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB\nregistry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB\nregistry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB\nregistry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB\nregistry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB\nregistry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB\nregistry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB\nregistry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB\nregistry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB\nregistry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB\nregistry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB\nregistry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB\nregistry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver 9.5.0 5.0.0\n\n## Containers @ testbed-node-1\n\n+ [[ 1 -eq -1 ]]\n+ echo\n+ echo '## Containers @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 ps\nCONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES\n5595f7aa7bde   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor\ne70175a7f54f   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api\n3947fc72dc2b   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   6 minutes ago       Up 6 minutes                           grafana\n79d75ef2089b   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter\n7bfe76f30932   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor\nc528ae2db2f9   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_memcached_exporter\n34200a94a9bb   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter\na619c7d8b861   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter\n31eed0c039c6   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share\nd7c831ca33ae   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_scheduler\nac2ea9870602   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data\na5e88825bb8f   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api\n7e841d0eab1a   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier\n281374c8a420   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener\nc732d68cf8a0   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator\nc34a9fd30b59   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api\n8438fd6d7aab   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central\n30fc5e345a33   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification\n6eb7797a56dd   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_worker\n0fa943cb70b5   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping\n9ccee7b0750c   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager\nff30b926a413   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent\ndbbf2a638521   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api\nab634d007474   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker\n80b69c1f45f8   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns\n0c654d6208e2   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer\nc9a047b3172f   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central\nc3785a6cb03e   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api\n8f6b5ce65f81   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9\nb5df48747341   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker\n97c106d5b111   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_keystone_listener\nb695ff5dbfeb   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api\n3740fc0af171   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup\n6e071ad5926e   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 31 minutes (healthy)                cinder_volume\ne794b78209c8   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler\nacd586036117   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api\ne56028a9ebc8   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api\n06f4dc20a799   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console\nb30b39ae1f62   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver\n12aa037e6068   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon\n5905244ae562   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy\n3f86776bbdd0   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor\n39ee1db54373   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   45 minutes ago      Up 44 minutes (healthy)                nova_api\n17d85da2dc3e   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler\n8ecb795d537e   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server\n2c6fcd0fb49d   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   53 minutes ago      Up 53 minutes (healthy)                placement_api\n2f709020f008   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone\neded742115a7   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet\n6c890b259434   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh\nef5500dbe92f   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-1\n4f9927e71ae2   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-1\n1fb5a820b9f6   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-1\n7afd57013429   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd\neb26e25add2e   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db\n4474d5901cb6   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db\nf3366fda0db6   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller\n8eea01b43193   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd\nfd6d0d5f070c   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db\ncbec62d736b0   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq\na2c9b8c73ba7   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb\nde191e12b2ba   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel\nf4bf64deb003   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis\n04c02bb470b8   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached\nc5a6b7eb3732   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             opensearch_dashboards\n16973c91b145   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch\n88a35f9e4664   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived\nbf5b20e3ad39   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql\na55c8b3a0c88   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy\nb85bba8a226d   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron\nbfa74148a0c1   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox\n0f65f0407b45   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd\n\n## Images @ testbed-node-1\n\n+ echo\n+ echo '## Images @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 images\nREPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE\nregistry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB\nregistry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB\nregistry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB\nregistry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB\nregistry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB\nregistry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB\nregistry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB\nregistry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB\nregistry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB\nregistry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB\nregistry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB\nregistry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB\nregistry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB\nregistry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB\nregistry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB\nregistry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB\nregistry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB\nregistry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB\nregistry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB\nregistry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB\nregistry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB\nregistry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB\nregistry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB\nregistry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB\nregistry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB\nregistry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB\nregistry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB\nregistry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB\nregistry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB\nregistry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB\nregistry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB\nregistry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB\nregistry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB\nregistry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB\nregistry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB\nregistry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB\nregistry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB\nregistry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB\nregistry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB\nregistry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB\nregistry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB\nregistry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver 9.5.0 5.0.0\n\n## Containers @ testbed-node-2\n\n+ [[ 1 -eq -1 ]]\n+ echo\n+ echo '## Containers @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 ps\nCONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES\ncb5a2e0036ae   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor\nb1fc9f76bfe1   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api\n79bb3cfa27d7   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   6 minutes ago       Up 6 minutes                           grafana\n74f7548606e0   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter\n6ff2ddb90324   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor\n7541b568abd6   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_memcached_exporter\nc8e9e481f34b   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter\n91072cae9492   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter\ne9bb963e7218   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share\n12a9083a7490   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_scheduler\n4f365a87ab5f   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data\n42b0d560762f   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api\n4511ec114d0d   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier\n39c62c61e597   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener\n4eb2be969478   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator\n3dd95017f326   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api\n93d1c0dc0e86   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central\nc225d72455b3   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification\na01140bcf6bf   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_worker\n0056447f6cdc   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping\n23eb3d9e27c0   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager\n649fe0554c73   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent\n87e3f0068950   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api\ndc2d6c4aae9a   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker\ncd5dacfed1b8   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns\nc0e8307f4672   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer\n68bd368adfdc   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central\nc51595c11699   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api\n8dffa9e9b29d   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9\n6d0216ea6747   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker\n2ad9f138b268   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_keystone_listener\n61e3accb5af7   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api\n4c7ed3dcd5fc   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup\ncf0ff34a99c9   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_volume\n67e117d5e66e   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler\n70fa2260ff57   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api\n799f8b17e49d   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api\n1230e722ebb3   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console\n1b7da13de50d   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver\n8142e29746fa   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon\n13a1d439e229   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy\n3df27d3081d1   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor\n02286aff1530   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_api\n69a542a499a4   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler\n04863e81a670   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server\n943c7539f102   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   54 minutes ago      Up 54 minutes (healthy)                placement_api\n1795602436d0   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone\n390c789c5ca0   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet\n32eece189277   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh\n4373766b4740   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-2\nc7fe63b73697   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-2\n2a382ea60872   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-2\n09e52c0af1ac   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd\n5cd3e7218a58   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db\n6a6022c2623e   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db\n0e57fca16035   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller\nb5135bee0b88   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd\n2e496286f4d6   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db\n774e79736876   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq\na33e8c9a11ce   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb\ndc6520f0098a   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel\n620d722dcf31   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis\na26c32c37c05   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached\n5df26a02d819   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             opensearch_dashboards\na0ce1c2c8df2   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch\n35762f95d946   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived\n4a8d2c92eea7   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql\n6a2161fe9af1   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy\nc05e2866aaa1   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron\nd6f4b784a67c   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox\n7742d32707c2   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd\n\n## Images @ testbed-node-2\n\n+ echo\n+ echo '## Images @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 images\nREPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE\nregistry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB\nregistry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB\nregistry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB\nregistry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB\nregistry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB\nregistry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB\nregistry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB\nregistry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB\nregistry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB\nregistry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB\nregistry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB\nregistry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB\nregistry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB\nregistry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB\nregistry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB\nregistry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB\nregistry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB\nregistry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB\nregistry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB\nregistry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB\nregistry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB\nregistry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB\nregistry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB\nregistry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB\nregistry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB\nregistry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB\nregistry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB\nregistry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB\nregistry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB\nregistry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB\nregistry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB\nregistry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB\nregistry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB\nregistry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB\nregistry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB\nregistry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB\nregistry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB\nregistry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB\nregistry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB\nregistry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB\nregistry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB\nregistry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB\nregistry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB\nregistry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB\nregistry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB\nregistry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB\nregistry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB\nregistry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB\nregistry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB\nregistry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB\nregistry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB\nregistry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB\n+ sh -c /opt/configuration/scripts/check-services.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# Ceph status\n\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n+ echo\n+ echo '# Ceph status'\n+ echo\n+ ceph -s\n  cluster:\n    id:     11111111-1111-1111-1111-111111111111\n    health: HEALTH_OK\n \n  services:\n    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 71m)\n    mgr: testbed-node-1(active, since 58m), standbys: testbed-node-0, testbed-node-2\n    mds: 1/1 daemons up, 2 standby\n    osd: 6 osds: 6 up (since 67m), 6 in (since 68m)\n    rgw: 3 daemons active (3 hosts, 1 zones)\n \n  data:\n    volumes: 1/1 healthy\n    pools:   14 pools, 401 pgs\n    objects: 556 objects, 2.2 GiB\n    usage:   7.0 GiB used, 113 GiB / 120 GiB avail\n    pgs:     401 active+clean\n \n\n# Ceph versions\n\n+ echo\n+ echo '# Ceph versions'\n+ echo\n+ ceph versions\n{\n    \"mon\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"mgr\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"osd\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6\n    },\n    \"mds\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"rgw\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"overall\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18\n    }\n}\n\n# Ceph OSD tree\n\n+ echo\n+ echo '# Ceph OSD tree'\n+ echo\n+ ceph osd df tree\nID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              \n-1         0.11691         -  120 GiB  7.0 GiB  6.7 GiB    6 KiB  369 MiB  113 GiB  5.87  1.00    -          root default           \n-3         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-3\n 0    hdd  0.01949   1.00000   20 GiB  840 MiB  779 MiB    1 KiB   62 MiB   19 GiB  4.11  0.70  189      up          osd.0          \n 3    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.5 GiB    1 KiB   62 MiB   18 GiB  7.63  1.30  201      up          osd.3          \n-5         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-4\n 2    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.4 GiB    1 KiB   62 MiB   19 GiB  7.20  1.23  206      up          osd.2          \n 5    hdd  0.01949   1.00000   20 GiB  928 MiB  867 MiB    1 KiB   62 MiB   19 GiB  4.54  0.77  186      up          osd.5          \n-7         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-5\n 1    hdd  0.01949   1.00000   20 GiB  1.2 GiB  1.1 GiB    1 KiB   62 MiB   19 GiB  5.95  1.01  192      up          osd.1          \n 4    hdd  0.01949   1.00000   20 GiB  1.2 GiB  1.1 GiB    1 KiB   62 MiB   19 GiB  5.79  0.99  196      up          osd.4          \n                       TOTAL  120 GiB  7.0 GiB  6.7 GiB  9.3 KiB  369 MiB  113 GiB  5.87                                            \nMIN/MAX VAR: 0.70/1.30  STDDEV: 1.27\n\n# Ceph monitor status\n\n+ echo\n+ echo '# Ceph monitor status'\n+ echo\n+ ceph mon stat\ne1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2\n\n# Ceph quorum status\n\n+ echo\n+ echo '# Ceph quorum status'\n+ echo\n+ ceph quorum_status\n+ jq\n{\n  \"election_epoch\": 8,\n  \"quorum\": [\n    0,\n    1,\n    2\n  ],\n  \"quorum_names\": [\n    \"testbed-node-0\",\n    \"testbed-node-1\",\n    \"testbed-node-2\"\n  ],\n  \"quorum_leader_name\": \"testbed-node-0\",\n  \"quorum_age\": 4267,\n  \"features\": {\n    \"quorum_con\": \"4540138322906710015\",\n    \"quorum_mon\": [\n      \"kraken\",\n      \"luminous\",\n      \"mimic\",\n      \"osdmap-prune\",\n      \"nautilus\",\n      \"octopus\",\n      \"pacific\",\n      \"elector-pinging\",\n      \"quincy\",\n      \"reef\"\n    ]\n  },\n  \"monmap\": {\n    \"epoch\": 1,\n    \"fsid\": \"11111111-1111-1111-1111-111111111111\",\n    \"modified\": \"2026-03-26T02:53:06.287135Z\",\n    \"created\": \"2026-03-26T02:53:06.287135Z\",\n    \"min_mon_release\": 18,\n    \"min_mon_release_name\": \"reef\",\n    \"election_strategy\": 1,\n    \"disallowed_leaders: \": \"\",\n    \"stretch_mode\": false,\n    \"tiebreaker_mon\": \"\",\n    \"removed_ranks: \": \"\",\n    \"features\": {\n      \"persistent\": [\n        \"kraken\",\n        \"luminous\",\n        \"mimic\",\n        \"osdmap-prune\",\n        \"nautilus\",\n        \"octopus\",\n        \"pacific\",\n        \"elector-pinging\",\n        \"quincy\",\n        \"reef\"\n      ],\n      \"optional\": []\n    },\n    \"mons\": [\n      {\n        \"rank\": 0,\n        \"name\": \"testbed-node-0\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.10:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.10:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.10:6789/0\",\n        \"public_addr\": \"192.168.16.10:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 1,\n        \"name\": \"testbed-node-1\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.11:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.11:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.11:6789/0\",\n        \"public_addr\": \"192.168.16.11:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 2,\n        \"name\": \"testbed-node-2\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.12:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.12:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.12:6789/0\",\n        \"public_addr\": \"192.168.16.12:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      }\n    ]\n  }\n}\n\n# Ceph free space status\n+ echo\n+ echo '# Ceph free space status'\n\n+ echo\n+ ceph df\n--- RAW STORAGE ---\nCLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED\nhdd    120 GiB  113 GiB  7.0 GiB   7.0 GiB       5.87\nTOTAL  120 GiB  113 GiB  7.0 GiB   7.0 GiB       5.87\n \n--- POOLS ---\nPOOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL\n.mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB\ncephfs_data                 2   32      0 B        0      0 B      0     35 GiB\ncephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB\ndefault.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.control         6   32      0 B        8      0 B      0     35 GiB\ndefault.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB\ndefault.rgw.meta            8   32      0 B        0      0 B      0     35 GiB\n.rgw.root                   9   32  3.9 KiB        8   64 KiB      0     52 GiB\nbackups                    10   32     19 B        2   12 KiB      0     35 GiB\nvolumes                    11   32     19 B        2   12 KiB      0     35 GiB\nimages                     12   32  2.2 GiB      299  6.7 GiB   5.98     35 GiB\nmetrics                    13   32     19 B        2   12 KiB      0     35 GiB\nvms                        14   32     19 B        2   12 KiB      0     35 GiB\n++ semver 9.5.0 5.0.0\n+ [[ 1 -eq -1 ]]\n+ [[ ! -e /etc/redhat-release ]]\n+ osism apply facts\n\u001b[32m2026-03-26 04:04:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8207bc9-b917-4707-81f1-417025aac495 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:04:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a8207bc9-b917-4707-81f1-417025aac495 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nThursday 26 March 2026  04:04:38 +0000 (0:00:00.296)       0:00:00.296 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nThursday 26 March 2026  04:04:39 +0000 (0:00:01.129)       0:00:01.425 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  04:04:41 +0000 (0:00:01.360)       0:00:02.785 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nThursday 26 March 2026  04:04:46 +0000 (0:00:05.714)       0:00:08.500 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:04:47 +0000 (0:00:00.589)       0:00:09.089 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.71s\nosism.commons.facts : Copy fact files ----------------------------------- 1.36s\nosism.commons.facts : Create custom facts directory --------------------- 1.13s\nGather facts for all hosts ---------------------------------------------- 0.59s\n+ osism validate ceph-mons\n\nPLAY [Ceph validate mons] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nThursday 26 March 2026  04:05:04 +0000 (0:00:00.452)       0:00:00.452 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nThursday 26 March 2026  04:05:05 +0000 (0:00:00.853)       0:00:01.306 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nThursday 26 March 2026  04:05:06 +0000 (0:00:00.812)       0:00:02.118 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nThursday 26 March 2026  04:05:06 +0000 (0:00:00.115)       0:00:02.234 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nThursday 26 March 2026  04:05:06 +0000 (0:00:00.259)       0:00:02.494 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nThursday 26 March 2026  04:05:07 +0000 (0:00:01.012)       0:00:03.506 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nThursday 26 March 2026  04:05:08 +0000 (0:00:00.267)       0:00:03.774 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:05:08 +0000 (0:00:00.407)       0:00:04.181 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mon is not running] ********************\nThursday 26 March 2026  04:05:08 +0000 (0:00:00.299)       0:00:04.480 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mon is running] ************************\nThursday 26 March 2026  04:05:09 +0000 (0:00:00.289)       0:00:04.770 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:05:09 +0000 (0:00:00.426)       0:00:05.196 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:05:09 +0000 (0:00:00.225)       0:00:05.421 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:05:09 +0000 (0:00:00.233)       0:00:05.655 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.261)       0:00:05.916 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.068)       0:00:05.985 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.071)       0:00:06.056 ******** \n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.071)       0:00:06.127 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.225)       0:00:06.353 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Prepare quorum test vars] ************************************************\nThursday 26 March 2026  04:05:10 +0000 (0:00:00.234)       0:00:06.588 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get monmap info from one mon container] **********************************\nThursday 26 March 2026  04:05:11 +0000 (0:00:00.115)       0:00:06.703 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set quorum test data] ****************************************************\nThursday 26 March 2026  04:05:12 +0000 (0:00:01.639)       0:00:08.343 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail quorum test if not all monitors are in quorum] **********************\nThursday 26 March 2026  04:05:13 +0000 (0:00:00.528)       0:00:08.871 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass quorum test if all monitors are in quorum] **************************\nThursday 26 March 2026  04:05:13 +0000 (0:00:00.133)       0:00:09.004 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set fsid test vars] ******************************************************\nThursday 26 March 2026  04:05:13 +0000 (0:00:00.333)       0:00:09.338 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail Cluster FSID test if FSID does not match configuration] *************\nThursday 26 March 2026  04:05:13 +0000 (0:00:00.320)       0:00:09.659 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass Cluster FSID test if it matches configuration] **********************\nThursday 26 March 2026  04:05:14 +0000 (0:00:00.128)       0:00:09.787 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare status test vars] ************************************************\nThursday 26 March 2026  04:05:14 +0000 (0:00:00.116)       0:00:09.904 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather status data] ******************************************************\nThursday 26 March 2026  04:05:14 +0000 (0:00:00.168)       0:00:10.073 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set health test data] ****************************************************\nThursday 26 March 2026  04:05:15 +0000 (0:00:01.416)       0:00:11.490 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable] *************************\nThursday 26 March 2026  04:05:16 +0000 (0:00:00.313)       0:00:11.803 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if health is acceptable] *****************************\nThursday 26 March 2026  04:05:16 +0000 (0:00:00.143)       0:00:11.947 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable (strict)] ****************\nThursday 26 March 2026  04:05:16 +0000 (0:00:00.165)       0:00:12.112 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if status is OK (strict)] ****************************\nThursday 26 March 2026  04:05:16 +0000 (0:00:00.142)       0:00:12.255 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nThursday 26 March 2026  04:05:16 +0000 (0:00:00.384)       0:00:12.640 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nThursday 26 March 2026  04:05:17 +0000 (0:00:00.310)       0:00:12.950 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:05:17 +0000 (0:00:00.281)       0:00:13.231 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:05:19 +0000 (0:00:02.146)       0:00:15.378 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:05:19 +0000 (0:00:00.278)       0:00:15.657 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:20 +0000 (0:00:00.270)       0:00:15.927 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:20 +0000 (0:00:00.079)       0:00:16.007 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:20 +0000 (0:00:00.100)       0:00:16.107 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nThursday 26 March 2026  04:05:20 +0000 (0:00:00.084)       0:00:16.191 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:05:22 +0000 (0:00:01.618)       0:00:17.809 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-26T04:05:05+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:05:23 +0000 (0:00:00.917)       0:00:18.727 ******** \n=============================================================================== \nAggregate test results step one ----------------------------------------- 2.15s\nGet monmap info from one mon container ---------------------------------- 1.64s\nWrite report file ------------------------------------------------------- 1.62s\nGather status data ------------------------------------------------------ 1.42s\nGet container info ------------------------------------------------------ 1.01s\nPrint report file information ------------------------------------------- 0.92s\nGet timestamp for report file ------------------------------------------- 0.85s\nCreate report output directory ------------------------------------------ 0.81s\nSet quorum test data ---------------------------------------------------- 0.53s\nSet test result to passed if ceph-mon is running ------------------------ 0.43s\nSet test result to passed if container is existing ---------------------- 0.41s\nPass cluster-health if status is OK (strict) ---------------------------- 0.38s\nPass quorum test if all monitors are in quorum -------------------------- 0.33s\nSet fsid test vars ------------------------------------------------------ 0.32s\nSet health test data ---------------------------------------------------- 0.31s\nSet validation result to passed if no test failed ----------------------- 0.31s\nPrepare test data ------------------------------------------------------- 0.30s\nSet test result to failed if ceph-mon is not running -------------------- 0.29s\nSet validation result to failed if a test failed ------------------------ 0.28s\nAggregate test results step two ----------------------------------------- 0.28s\n+ osism validate ceph-mgrs\n\nPLAY [Ceph validate mgrs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nThursday 26 March 2026  04:05:40 +0000 (0:00:00.468)       0:00:00.468 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nThursday 26 March 2026  04:05:41 +0000 (0:00:00.832)       0:00:01.301 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nThursday 26 March 2026  04:05:42 +0000 (0:00:00.980)       0:00:02.281 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nThursday 26 March 2026  04:05:42 +0000 (0:00:00.146)       0:00:02.427 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nThursday 26 March 2026  04:05:42 +0000 (0:00:00.311)       0:00:02.739 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nThursday 26 March 2026  04:05:43 +0000 (0:00:01.008)       0:00:03.748 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nThursday 26 March 2026  04:05:44 +0000 (0:00:00.315)       0:00:04.064 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:05:44 +0000 (0:00:00.539)       0:00:04.603 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mgr is not running] ********************\nThursday 26 March 2026  04:05:44 +0000 (0:00:00.340)       0:00:04.943 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mgr is running] ************************\nThursday 26 March 2026  04:05:45 +0000 (0:00:00.299)       0:00:05.242 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:05:45 +0000 (0:00:00.544)       0:00:05.787 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.247)       0:00:06.035 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.249)       0:00:06.285 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.246)       0:00:06.532 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.070)       0:00:06.602 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.071)       0:00:06.674 ******** \n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.075)       0:00:06.750 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nThursday 26 March 2026  04:05:46 +0000 (0:00:00.248)       0:00:06.999 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Define mgr module test vars] *********************************************\nThursday 26 March 2026  04:05:47 +0000 (0:00:00.245)       0:00:07.245 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather list of mgr modules] **********************************************\nThursday 26 March 2026  04:05:47 +0000 (0:00:00.118)       0:00:07.363 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Parse mgr module list from json] *****************************************\nThursday 26 March 2026  04:05:49 +0000 (0:00:02.082)       0:00:09.446 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Extract list of enabled mgr modules] *************************************\nThursday 26 March 2026  04:05:49 +0000 (0:00:00.462)       0:00:09.908 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail test if mgr modules are disabled that should be enabled] ************\nThursday 26 March 2026  04:05:50 +0000 (0:00:00.323)       0:00:10.231 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass test if required mgr modules are enabled] ***************************\nThursday 26 March 2026  04:05:50 +0000 (0:00:00.194)       0:00:10.425 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nThursday 26 March 2026  04:05:50 +0000 (0:00:00.143)       0:00:10.569 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nThursday 26 March 2026  04:05:50 +0000 (0:00:00.263)       0:00:10.832 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:05:51 +0000 (0:00:00.254)       0:00:11.087 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:05:52 +0000 (0:00:01.280)       0:00:12.368 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:05:52 +0000 (0:00:00.256)       0:00:12.625 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:52 +0000 (0:00:00.254)       0:00:12.879 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:52 +0000 (0:00:00.071)       0:00:12.951 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:05:53 +0000 (0:00:00.091)       0:00:13.042 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nThursday 26 March 2026  04:05:53 +0000 (0:00:00.270)       0:00:13.313 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:05:54 +0000 (0:00:01.330)       0:00:14.644 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-26T04:05:41+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:05:55 +0000 (0:00:00.405)       0:00:15.049 ******** \n=============================================================================== \nGather list of mgr modules ---------------------------------------------- 2.08s\nWrite report file ------------------------------------------------------- 1.33s\nAggregate test results step one ----------------------------------------- 1.28s\nGet container info ------------------------------------------------------ 1.01s\nCreate report output directory ------------------------------------------ 0.98s\nGet timestamp for report file ------------------------------------------- 0.83s\nSet test result to passed if ceph-mgr is running ------------------------ 0.54s\nSet test result to passed if container is existing ---------------------- 0.54s\nParse mgr module list from json ----------------------------------------- 0.46s\nFlush handlers ---------------------------------------------------------- 0.43s\nPrint report file information ------------------------------------------- 0.41s\nPrepare test data ------------------------------------------------------- 0.34s\nExtract list of enabled mgr modules ------------------------------------- 0.32s\nSet test result to failed if container is missing ----------------------- 0.32s\nPrepare test data for container existance test -------------------------- 0.31s\nSet test result to failed if ceph-mgr is not running -------------------- 0.30s\nSet validation result to passed if no test failed ----------------------- 0.26s\nAggregate test results step two ----------------------------------------- 0.26s\nSet validation result to failed if a test failed ------------------------ 0.25s\nAggregate test results step three --------------------------------------- 0.25s\n+ osism validate ceph-osds\n\nPLAY [Ceph validate OSDs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nThursday 26 March 2026  04:06:12 +0000 (0:00:00.441)       0:00:00.441 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nThursday 26 March 2026  04:06:13 +0000 (0:00:00.872)       0:00:01.314 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nThursday 26 March 2026  04:06:14 +0000 (0:00:00.554)       0:00:01.869 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nThursday 26 March 2026  04:06:14 +0000 (0:00:00.789)       0:00:02.658 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nThursday 26 March 2026  04:06:14 +0000 (0:00:00.133)       0:00:02.792 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nThursday 26 March 2026  04:06:15 +0000 (0:00:00.123)       0:00:02.916 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nThursday 26 March 2026  04:06:15 +0000 (0:00:00.312)       0:00:03.228 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nThursday 26 March 2026  04:06:15 +0000 (0:00:00.161)       0:00:03.390 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate total number of OSDs in cluster] *******************************\nThursday 26 March 2026  04:06:15 +0000 (0:00:00.315)       0:00:03.705 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:06:16 +0000 (0:00:00.784)       0:00:04.490 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers on host] *********************************\nThursday 26 March 2026  04:06:16 +0000 (0:00:00.292)       0:00:04.782 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '6a47ae157c609fdba1c3cb78cfabe35faae1300906bbf5c42bf9812cf5ae3fc3', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '19873f576682f61766adcb2a0a3bd7e7d79b6aaf7f359db744e235b802241d18', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '64dc82767cbb91bfba3d4fbe0f910f1c1b95d0856e1921a113524e50b04683ae', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '905865d8832bd6a48e47e77f25a307fabc6f4262f500433c693482e257e98351', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0a2ea494cefa3bc62f3a02d275955bf06259547cc9c9e7d9a6af1873fa1e568c', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'eedb79789b5f6a192a1b22b4016b1979f5d3369e8b3a5882bd980f73e378d679', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '99dc660051ff8aa712dc0eb018d8040e264e7abe5fee8210dbf2a9b2887da177', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b7b3e024c8e1cc2e99343381e2b424833301da2438f00dc4ab4f87d51ae0ce3d', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '261fb7fe11f046be78125f2f61829982321ff86ca846ffb773a8692a92cdc390', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a3e70f4017c0ded81315b1398083acb671abe365698c575bfc6fc4c9e4d55223', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'eacf9bcf824e7e1d8ff361b32cadee04f8af28293f8b9c40a61bc7a939d2150b', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': 'dd1dcbbb5855901401655533a58e1a34834b99fe4c43bf30a9fcba2126bd3e51', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': '30b29e57211792d4ff597f247627c18d4ebc0c7ea1e4f1feaa4a30caa4ab2f31', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ce801b2b1273df80d38b32bfc814e9fae0bff5a3f34069b7cd0581325b281bfc', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'c319b5f6aba28f04e9d72db4d66c7ea078f78b2519ee091016ef2335532f1d29', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a8c7c868a3e502d499713dbb69d7049568a671e189c4d31df99d14f7de70569d', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ebb9b42d163cdb490d062e6074b99691af113a221b62481074d5a150e8044129', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '2c477485e8286e4a0e2e6d3499c0dc253c6c858a2d7106f81cae62bb3c374eea', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'f2f4183c1750edb2a41639c237d8eb38ee30d96ee3593749616ba4f4b9c5c515', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e18421415e4ffa13dbe3a848dfdf8b91a1323c1f464879f5ec2e6aab512a0f6c', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd775a0281c84c1125a58c04446db7e74566793216c3126ce998f8163abb8bc89', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '9ef2f3ed84c16fec0eb076bb311683c068143380e4f67d2cb7d7ab8c906fed0b', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3ac4c9380a953c178f1b4d8a9ebd52716be1370e8865b0c2c7891f740028771a', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '281009e18ac6262f07d27fc06505b3524e5a3cb883aa16c0504e87f326290c63', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ac7b1afd4a8f6ca6f58ac6dc079c6e531e13b8a15c2c277555ce9c5de25c2d64', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '5b1338247b29769164d10e260f095122dca1f89297144212768f1fd6c4fb6fb7', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ff5b1b914ed4c437466f8660c32a9afbeb3695e9ac20e478f9c1ce409f1ed09e', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'bdc4ceb66ebf8c37ba5c65183f623ca08c64fe8018a102a4f0f5eabea3c15fc2', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e79d34b43243c4ec6976fac2d1089c8f2f3486f55094e31b12c811ad29a25c1c', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f3218cdcc9e6ffb5ab7cbc4e409a97c5cba656cef510c1e994aae98145b31869', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '49e8eee6f6a9043366b38d966aa36fed7ffa7a7c6fc5094b528cb58857b99f0f', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': 'c2bbcab38cdb8bcb221ca845f7476b10094b0cf3cf85313738dd3864b6053f52', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '612bc1ad6c697915f5f79c2f9ad3754ced490c89063b9bfdddf3dc5ed1737430', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '2314752bd091bb01a3c9c9f4d8f863e576f800a689d5703521d8306adffd76e2', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0a9ebc944b2de755c4512d2b1442597d5fe4b5297d62fadd6fd92902e760637f', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd155fbae7e1e3e7432c43233b801b4ce10005b36cfb24888760f62814956a602', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'abe890022cb2e1a31afca9ad0fc40511ef374822a7fcd357597aec36b3b503d0', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '5a0b086f92fe22dde447751035ee4785535bb4d9089ba71717434afd60163dc7', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'ea7350e3ce894a85dd937116b86ff58eefc778996eb3fc48d232d1c1dac27def', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'da395bf52b74e3fda92ab38bc0ddabb7e29028090e8635bdc44d7151a273550e', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7d2f8364b0ceaf10c88bb844de62d071fc4f0a0c28a054f6da0b22c9d76e89c3', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7a175bd7fe74190008d3ab3b9bbb0e7d06a5c2d88606abf1f5094899ee6eeadb', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9c0845a7a0f6ebf9eaf9aa84f6ffbeab245c4ade1433d91304b63e3259f88214', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b7e86c8b6f142da7eb27605a4d20d1731392d015424cc821aa9ce2b9dfb09a49', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f7d648df1e97280d44e097087185c63d2d79f38b50b281d365f017a81c42fc34', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8046f86788dee700330fed8eddfd693504584dd39873e897efa033060d800080', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b2877433cc1e735f486090df103fc266a4811a2c8e87200a6ec7cb1adbcafcc4', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'a31876dbc950b24a0d90116efabb4b484220e273b36b95211ea490ba31cbf16f', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c7485c82d2f1d547f6bd4f7b5f77839137922770f8e90166964bc693f5838c04', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '53b57430f3c2ea3e26a0313bd0f570ea9f96ee2cccf3420967739a15176d48d7', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '89e6fe0b290936f07e5cd92c26a65d7e4bad4aab8803b8ba0f073aa6f33518d8', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'da50a7d28d968683ce4d4ca44f0cc1ebfe8cb21bf5813a11fffb890dc0b60b07', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '837bfa57508f3be2114e78e4ad24b6ea1a5bad3c5b3d7e9b2025c2367a5971e5', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4d99cd8c4e4111585315968fcf3c1d2f5e140741723d97a5b22057ceeb981db5', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '30b368a03fb8c09907c752cec0533ecbd1ff372b353299d4030ee6f65a37b256', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'a6edd44b74454d1b03eb0b9925de800df0e9da371df15c735dedb32ddf11cc06', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'af8e0c0cd5013d2851f9115b1362788481251f73a98ac8f2abdd4398c7c1a67b', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m\n\nTASK [Get count of ceph-osd containers on host] ********************************\nThursday 26 March 2026  04:06:17 +0000 (0:00:00.515)       0:00:05.298 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed when count of containers is wrong] *************\nThursday 26 March 2026  04:06:17 +0000 (0:00:00.296)       0:00:05.594 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if count matches] ******************************\nThursday 26 March 2026  04:06:18 +0000 (0:00:00.481)       0:00:06.075 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:06:18 +0000 (0:00:00.313)       0:00:06.389 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers that are not running] ********************\nThursday 26 March 2026  04:06:18 +0000 (0:00:00.304)       0:00:06.693 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get count of ceph-osd containers that are not running] *******************\nThursday 26 March 2026  04:06:19 +0000 (0:00:00.326)       0:00:07.020 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nThursday 26 March 2026  04:06:19 +0000 (0:00:00.504)       0:00:07.524 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nThursday 26 March 2026  04:06:19 +0000 (0:00:00.304)       0:00:07.828 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if all containers are running] *****************\nThursday 26 March 2026  04:06:20 +0000 (0:00:00.307)       0:00:08.136 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:06:20 +0000 (0:00:00.336)       0:00:08.472 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:06:21 +0000 (0:00:00.680)       0:00:09.153 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:06:21 +0000 (0:00:00.275)       0:00:09.428 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:21 +0000 (0:00:00.255)       0:00:09.683 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:21 +0000 (0:00:00.068)       0:00:09.752 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:21 +0000 (0:00:00.069)       0:00:09.821 ******** \n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:06:22 +0000 (0:00:00.079)       0:00:09.901 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail early due to containers not running] ********************************\nThursday 26 March 2026  04:06:22 +0000 (0:00:00.244)       0:00:10.145 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:06:22 +0000 (0:00:00.248)       0:00:10.394 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set _mon_hostname fact] **************************************************\nThursday 26 March 2026  04:06:22 +0000 (0:00:00.286)       0:00:10.680 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get ceph osd tree] *******************************************************\nThursday 26 March 2026  04:06:23 +0000 (0:00:00.679)       0:00:11.360 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Parse osd tree from JSON] ************************************************\nThursday 26 March 2026  04:06:25 +0000 (0:00:01.722)       0:00:13.082 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get OSDs that are not up or in] ******************************************\nThursday 26 March 2026  04:06:25 +0000 (0:00:00.136)       0:00:13.219 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Fail test if OSDs are not up or in] **************************************\nThursday 26 March 2026  04:06:25 +0000 (0:00:00.318)       0:00:13.538 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Pass test if OSDs are all up and in] *************************************\nThursday 26 March 2026  04:06:25 +0000 (0:00:00.130)       0:00:13.668 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:06:25 +0000 (0:00:00.131)       0:00:13.800 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [List ceph LVM volumes and collect data] **********************************\nThursday 26 March 2026  04:06:26 +0000 (0:00:00.303)       0:00:14.103 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Parse LVM data as JSON] **************************************************\nThursday 26 March 2026  04:06:28 +0000 (0:00:02.574)       0:00:16.678 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get unencrypted and encrypted OSDs] **************************************\nThursday 26 March 2026  04:06:29 +0000 (0:00:00.326)       0:00:17.004 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of encrypted OSDs does not match] **************************\nThursday 26 March 2026  04:06:29 +0000 (0:00:00.581)       0:00:17.586 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of encrypted OSDs equals count of OSDs] ********************\nThursday 26 March 2026  04:06:30 +0000 (0:00:00.306)       0:00:17.892 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of unencrypted OSDs does not match] ************************\nThursday 26 March 2026  04:06:30 +0000 (0:00:00.565)       0:00:18.458 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************\nThursday 26 March 2026  04:06:30 +0000 (0:00:00.311)       0:00:18.769 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nThursday 26 March 2026  04:06:31 +0000 (0:00:00.297)       0:00:19.067 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get CRUSH node data of each OSD host and root node childs] ***************\nThursday 26 March 2026  04:06:31 +0000 (0:00:00.503)       0:00:19.571 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate sub test expression results] ***********************************\nThursday 26 March 2026  04:06:32 +0000 (0:00:00.776)       0:00:20.347 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail test if any sub test failed] ****************************************\nThursday 26 March 2026  04:06:32 +0000 (0:00:00.330)       0:00:20.677 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass test if no sub test failed] *****************************************\nThursday 26 March 2026  04:06:33 +0000 (0:00:00.339)       0:00:21.017 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nThursday 26 March 2026  04:06:33 +0000 (0:00:00.507)       0:00:21.525 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nThursday 26 March 2026  04:06:33 +0000 (0:00:00.283)       0:00:21.808 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nThursday 26 March 2026  04:06:34 +0000 (0:00:00.264)       0:00:22.073 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nThursday 26 March 2026  04:06:35 +0000 (0:00:01.680)       0:00:23.753 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nThursday 26 March 2026  04:06:36 +0000 (0:00:00.276)       0:00:24.030 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:36 +0000 (0:00:00.248)       0:00:24.279 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:36 +0000 (0:00:00.070)       0:00:24.349 ******** \n\nTASK [Flush handlers] **********************************************************\nThursday 26 March 2026  04:06:36 +0000 (0:00:00.070)       0:00:24.419 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nThursday 26 March 2026  04:06:36 +0000 (0:00:00.074)       0:00:24.494 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nThursday 26 March 2026  04:06:38 +0000 (0:00:01.557)       0:00:26.051 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-26T04:06:13+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:06:38 +0000 (0:00:00.590)       0:00:26.642 ******** \n=============================================================================== \nList ceph LVM volumes and collect data ---------------------------------- 2.57s\nGet ceph osd tree ------------------------------------------------------- 1.72s\nAggregate test results step one ----------------------------------------- 1.68s\nWrite report file ------------------------------------------------------- 1.56s\nGet timestamp for report file ------------------------------------------- 0.87s\nCreate report output directory ------------------------------------------ 0.79s\nCalculate total number of OSDs in cluster ------------------------------- 0.78s\nGet CRUSH node data of each OSD host and root node childs --------------- 0.78s\nAggregate test results step one ----------------------------------------- 0.68s\nSet _mon_hostname fact -------------------------------------------------- 0.68s\nPrint report file information ------------------------------------------- 0.59s\nGet unencrypted and encrypted OSDs -------------------------------------- 0.58s\nPass if count of encrypted OSDs equals count of OSDs -------------------- 0.57s\nGet extra vars for Ceph configuration ----------------------------------- 0.55s\nGet list of ceph-osd containers on host --------------------------------- 0.52s\nPass test if no sub test failed ----------------------------------------- 0.51s\nGet count of ceph-osd containers that are not running ------------------- 0.50s\nPrepare test data ------------------------------------------------------- 0.50s\nSet test result to failed when count of containers is wrong ------------- 0.48s\nFail test if any sub test failed ---------------------------------------- 0.34s\n+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ -e /etc/redhat-release ]]\n+ source /etc/os-release\n++ PRETTY_NAME='Ubuntu 24.04.4 LTS'\n++ NAME=Ubuntu\n++ VERSION_ID=24.04\n++ VERSION='24.04.4 LTS (Noble Numbat)'\n++ VERSION_CODENAME=noble\n++ ID=ubuntu\n++ ID_LIKE=debian\n++ HOME_URL=https://www.ubuntu.com/\n++ SUPPORT_URL=https://help.ubuntu.com/\n++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/\n++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\n++ UBUNTU_CODENAME=noble\n++ LOGO=ubuntu-logo\n+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]\n+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'\n+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n\n# Status of Elasticsearch\n\n+ pushd /opt/configuration/contrib\n+ echo\n+ echo '# Status of Elasticsearch'\n+ echo\n+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s\nOK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0\n\n# Status of MariaDB\n\n+ echo\n+ echo '# Status of MariaDB'\n+ echo\n++ semver 9.5.0 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]\n+ MARIADB_USER=root_shard_0\n+ bash nagios-plugins/check_galera_cluster -u root_shard_0 -p password -H api-int.testbed.osism.xyz -c 1\nReading package lists...\nBuilding dependency tree...\nReading state information...\nbc is already the newest version (1.07.1-3ubuntu4).\nbc set to manually installed.\n0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.\nOK: number of NODES = 3 (wsrep_cluster_size)\n\n# Status of Prometheus\n\n+ echo\n+ echo '# Status of Prometheus'\n+ echo\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy\nUnauthorized\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready\nUnauthorized\n\n# Status of RabbitMQ\n\n+ echo\n+ echo '# Status of RabbitMQ'\n+ echo\n++ semver 9.5.0 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]\n+ perl nagios-plugins/check_rabbitmq_cluster --ssl 1 -H api-int.testbed.osism.xyz -u openstack -p password\nRABBITMQ_CLUSTER OK - nb_running_node OK (3) nb_running_disc_node OK (3) nb_running_ram_node OK (0)\n\n# Status of Redis\n\n+ echo\n+ echo '# Status of Redis'\n+ echo\n+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j\nTCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.002068s;;;0.000000;10.000000\n\n+ popd\n+ echo\n# Create backup of MariaDB database\n\n+ echo '# Create backup of MariaDB database'\n+ echo\n+ osism apply mariadb_backup -e mariadb_backup_type=full\n\u001b[32m2026-03-26 04:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7b1aa3d5-846c-4962-a0bc-63f2776dc6a9 (mariadb_backup) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7b1aa3d5-846c-4962-a0bc-63f2776dc6a9 (mariadb_backup) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:07:11 +0000 (0:00:00.177)       0:00:00.177 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:07:11 +0000 (0:00:00.313)       0:00:00.491 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nThursday 26 March 2026  04:07:12 +0000 (0:00:00.652)       0:00:01.144 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  04:07:12 +0000 (0:00:00.429)       0:00:01.574 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Get MariaDB container facts] ***********************************\nThursday 26 March 2026  04:07:13 +0000 (0:00:00.572)       0:00:02.147 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Taking full database backup via Mariabackup] *******************\nThursday 26 March 2026  04:07:16 +0000 (0:00:03.248)       0:00:05.395 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nSTILL ALIVE [task 'mariadb : Taking full database backup via Mariabackup' is running] ***\n\nSTILL ALIVE [task 'mariadb : Taking full database backup via Mariabackup' is running] ***\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nThursday 26 March 2026  04:10:25 +0000 (0:03:08.742)       0:03:14.138 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nThursday 26 March 2026  04:10:25 +0000 (0:00:00.304)       0:03:14.442 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:10:26 +0000 (0:00:00.435)       0:03:14.878 ******** \n=============================================================================== \nmariadb : Taking full database backup via Mariabackup ----------------- 188.74s\nmariadb : Get MariaDB container facts ----------------------------------- 3.25s\nGroup hosts based on enabled services ----------------------------------- 0.65s\nmariadb : include_tasks ------------------------------------------------- 0.57s\nInclude mariadb post-upgrade.yml ---------------------------------------- 0.44s\nmariadb : Group MariaDB hosts based on shards --------------------------- 0.43s\nGroup hosts based on Kolla action --------------------------------------- 0.31s\nInclude mariadb post-deploy.yml ----------------------------------------- 0.30s\n+ sh -c /opt/configuration/scripts/check/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# OpenStack endpoints\n\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n+ export OS_CLOUD=admin\n+ OS_CLOUD=admin\n+ echo\n+ echo '# OpenStack endpoints'\n+ echo\n+ openstack endpoint list\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| 0033fc892da9412fa8a093fd5d2a963f | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |\n| 08783899140940f1ab1e28d0eb212655 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |\n| 0a595fefe7e94bd7a92b4bec73d647a6 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |\n| 186cae91475646dfbbc932451c78d918 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |\n| 2070bf277fa04949b94df4bb73a01ad5 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |\n| 214d96bfc5c74deaa0287323a323b818 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |\n| 3089431d8d8b4f0ba7824a5faf669f60 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |\n| 35fc90d9137b4d7f92d7ac68014b9fc6 | RegionOne | manila       | share           | True    | internal  | https://api-int.testbed.osism.xyz:8786/v1/%(tenant_id)s             |\n| 3b15fd0003e04f9f8aa0c22a2ea26c73 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |\n| 4fbec3dc7ce74b8cb3c6d0974f2e9e1d | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |\n| 52db55a577c242bea758b13e5034c75a | RegionOne | aodh         | alarming        | True    | public    | https://api.testbed.osism.xyz:8042                                  |\n| 56d8259ac37946e2aff58d5495dc0f81 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |\n| 57fa80cce7544a10a586776995a291a6 | RegionOne | skyline      | panel           | True    | internal  | https://api-int.testbed.osism.xyz:9998                              |\n| 75c163e497a14d5bb1f2b9ee146083e9 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |\n| 856b24feafb3443b9aaeae1f19b86647 | RegionOne | manilav2     | sharev2         | True    | internal  | https://api-int.testbed.osism.xyz:8786/v2                           |\n| 8ba59f40ae3b4193b5819b7c08fd3243 | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |\n| 9809bb157e744c398e8f9c7fcac26a1a | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |\n| 9989f77bad6840df9a22ad29181f93ea | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |\n| a59948b5184e45e084d6aaa31149f210 | RegionOne | skyline      | panel           | True    | public    | https://api.testbed.osism.xyz:9998                                  |\n| a5d9ce36abcb47c3b75e5988e8e015df | RegionOne | manilav2     | sharev2         | True    | public    | https://api.testbed.osism.xyz:8786/v2                               |\n| b53ac5aaf5914c0cb65e717d7642ad90 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |\n| c577ae50a66d49879af93d9946c176b8 | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |\n| c8cf89a51f61454891009bb19cafd4c4 | RegionOne | aodh         | alarming        | True    | internal  | https://api-int.testbed.osism.xyz:8042                              |\n| ce290ff5801e446f8041086f7a131bab | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |\n| d4f79ca24681457c8ddb3d19e9973a63 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |\n| db3adab6cd89404792a59d0f59461602 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |\n| db945c2decd04c0bb50c44206ee77a79 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |\n| e3fa99a3313a44d1acb8aa7427f045de | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |\n| f1d018f3e5f6498c9420138b7ade2a36 | RegionOne | manila       | share           | True    | public    | https://api.testbed.osism.xyz:8786/v1/%(tenant_id)s                 |\n| f7da6a40e78f462ea2d8069b207f9e0d | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n\n# Cinder\n\n+ echo\n+ echo '# Cinder'\n+ echo\n+ openstack volume service list\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| Binary           | Host                       | Zone     | Status  | State | Updated At                 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |\n| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |\n| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |\n| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:24.000000 |\n| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:31.000000 |\n| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:32.000000 |\n| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-26T04:10:29.000000 |\n| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-26T04:10:31.000000 |\n| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-26T04:10:32.000000 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n\n# Neutron\n\n+ echo\n+ echo '# Neutron'\n+ echo\n+ openstack network agent list\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |\n| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n+ openstack network service provider list\n+---------------+------+---------+\n| Service Type  | Name | Default |\n+---------------+------+---------+\n| L3_ROUTER_NAT | ovn  | True    |\n+---------------+------+---------+\n\n# Nova\n\n+ echo\n+ echo '# Nova'\n+ echo\n+ openstack compute service list\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| 86eb2092-fe79-4181-8a5e-f8a408c1e3b9 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-26T04:10:40.000000 |\n| dd2eeb5e-c5d1-428b-bbce-901a429360a4 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-26T04:10:35.000000 |\n| 5892028d-cfad-462c-9137-727335627c04 | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-26T04:10:37.000000 |\n| f48bec24-2526-47fb-89c7-f4540915a0fa | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-26T04:10:37.000000 |\n| 7a41d58a-be90-4c07-b6b6-e07900ee5ad4 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-26T04:10:39.000000 |\n| f0aa1fcd-e220-419e-887b-6a41fe3dd532 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-26T04:10:40.000000 |\n| 83b43651-bd92-4d18-a2b5-fcbdc87357c5 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |\n| b9e61882-f01c-4632-a3ea-b296d9feba1c | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |\n| eff6a403-1dbd-4531-b082-376def61e6f9 | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n+ openstack hypervisor list\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| 227f7a23-f93c-4b92-b429-5a83b73d9fd5 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |\n| 147427a8-4597-4a2c-823c-3005e03af22a | testbed-node-3      | QEMU            | 192.168.16.13 | up    |\n| 4a11d35e-b357-4267-8292-b9149dc34db0 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n+ echo\n\n# Run OpenStack test play\n\n+ echo '# Run OpenStack test play'\n+ echo\n+ osism apply --environment openstack test\n\u001b[32m2026-03-26 04:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m\n\u001b[32m2026-03-26 04:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 49017abc-3eae-4005-afad-1c43fa7ba55f (test) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 49017abc-3eae-4005-afad-1c43fa7ba55f (test) has been started and output is visible here.\u001b[0m\n\nPLAY [Create test project] *****************************************************\n\nTASK [Create test domain] ******************************************************\nThursday 26 March 2026  04:11:01 +0000 (0:00:00.080)       0:00:00.080 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test-admin user] **************************************************\nThursday 26 March 2026  04:11:04 +0000 (0:00:03.755)       0:00:03.835 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add manager role to user test-admin] *************************************\nThursday 26 March 2026  04:11:09 +0000 (0:00:04.139)       0:00:07.975 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test project] *****************************************************\nThursday 26 March 2026  04:11:15 +0000 (0:00:06.524)       0:00:14.499 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test user] ********************************************************\nThursday 26 March 2026  04:11:19 +0000 (0:00:03.974)       0:00:18.474 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add member roles to user test] *******************************************\nThursday 26 March 2026  04:11:23 +0000 (0:00:04.185)       0:00:22.659 ******** \n\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m\n\nTASK [Create test server group] ************************************************\nThursday 26 March 2026  04:11:35 +0000 (0:00:11.421)       0:00:34.080 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create ssh security group] ***********************************************\nThursday 26 March 2026  04:11:40 +0000 (0:00:05.096)       0:00:39.177 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to ssh security group] ******************************************\nThursday 26 March 2026  04:11:45 +0000 (0:00:04.925)       0:00:44.102 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create icmp security group] **********************************************\nThursday 26 March 2026  04:11:49 +0000 (0:00:04.320)       0:00:48.423 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to icmp security group] *****************************************\nThursday 26 March 2026  04:11:53 +0000 (0:00:03.923)       0:00:52.347 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test keypair] *****************************************************\nThursday 26 March 2026  04:11:57 +0000 (0:00:04.091)       0:00:56.438 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test network] *****************************************************\nThursday 26 March 2026  04:12:01 +0000 (0:00:03.894)       0:01:00.332 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test subnet] ******************************************************\nThursday 26 March 2026  04:12:06 +0000 (0:00:04.908)       0:01:05.241 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test router] ******************************************************\nThursday 26 March 2026  04:12:12 +0000 (0:00:05.809)       0:01:11.050 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Manage test instances and volumes] ***************************************\n\nTASK [Get test server group] ***************************************************\nThursday 26 March 2026  04:12:23 +0000 (0:00:11.547)       0:01:22.598 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Detach test volume] ******************************************************\nThursday 26 March 2026  04:12:27 +0000 (0:00:03.550)       0:01:26.149 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test volume] ******************************************************\nThursday 26 March 2026  04:12:27 +0000 (0:00:00.058)       0:01:26.207 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test instances] ***************************************************\nThursday 26 March 2026  04:12:27 +0000 (0:00:00.055)       0:01:26.263 ******** \n\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m\n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Wait for instance deletion to complete] **********************************\nThursday 26 March 2026  04:12:27 +0000 (0:00:00.157)       0:01:26.420 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Create test instances] ***************************************************\nThursday 26 March 2026  04:12:27 +0000 (0:00:00.152)       0:01:26.572 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for instance creation to complete] **********************************\nThursday 26 March 2026  04:12:32 +0000 (0:00:04.784)       0:01:31.357 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j538260717191.3753', 'results_file': '/ansible/.ansible_async/j538260717191.3753', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j899561745004.3778', 'results_file': '/ansible/.ansible_async/j899561745004.3778', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j588604499832.3803', 'results_file': '/ansible/.ansible_async/j588604499832.3803', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j405640357128.3828', 'results_file': '/ansible/.ansible_async/j405640357128.3828', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j525201259332.3853', 'results_file': '/ansible/.ansible_async/j525201259332.3853', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add metadata to instances] ***********************************************\nThursday 26 March 2026  04:13:29 +0000 (0:00:57.245)       0:02:28.602 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for metadata to be added] *******************************************\nThursday 26 March 2026  04:13:34 +0000 (0:00:04.390)       0:02:32.993 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j984712504400.3965', 'results_file': '/ansible/.ansible_async/j984712504400.3965', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j533774217712.3990', 'results_file': '/ansible/.ansible_async/j533774217712.3990', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j898183755290.4015', 'results_file': '/ansible/.ansible_async/j898183755290.4015', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j768463755450.4040', 'results_file': '/ansible/.ansible_async/j768463755450.4040', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j698557361688.4065', 'results_file': '/ansible/.ansible_async/j698557361688.4065', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add tag to instances] ****************************************************\nThursday 26 March 2026  04:13:43 +0000 (0:00:09.467)       0:02:42.461 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for tags to be added] ***********************************************\nThursday 26 March 2026  04:13:48 +0000 (0:00:04.613)       0:02:47.075 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j766959703281.4134', 'results_file': '/ansible/.ansible_async/j766959703281.4134', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j744976519201.4159', 'results_file': '/ansible/.ansible_async/j744976519201.4159', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j468878930547.4191', 'results_file': '/ansible/.ansible_async/j468878930547.4191', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j481909587613.4217', 'results_file': '/ansible/.ansible_async/j481909587613.4217', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j4746500476.4243', 'results_file': '/ansible/.ansible_async/j4746500476.4243', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Create test volume] ******************************************************\nThursday 26 March 2026  04:13:57 +0000 (0:00:09.731)       0:02:56.806 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Attach test volume] ******************************************************\nThursday 26 March 2026  04:14:04 +0000 (0:00:06.098)       0:03:02.904 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create floating ip address] **********************************************\nThursday 26 March 2026  04:14:17 +0000 (0:00:13.445)       0:03:16.350 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Print floating ip address] ***********************************************\nThursday 26 March 2026  04:14:22 +0000 (0:00:05.342)       0:03:21.693 ******** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"192.168.112.156\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:14:22 +0000 (0:00:00.050)       0:03:21.744 ******** \n=============================================================================== \nWait for instance creation to complete --------------------------------- 57.25s\nAttach test volume ----------------------------------------------------- 13.45s\nCreate test router ----------------------------------------------------- 11.55s\nAdd member roles to user test ------------------------------------------ 11.42s\nWait for tags to be added ----------------------------------------------- 9.73s\nWait for metadata to be added ------------------------------------------- 9.47s\nAdd manager role to user test-admin ------------------------------------- 6.52s\nCreate test volume ------------------------------------------------------ 6.10s\nCreate test subnet ------------------------------------------------------ 5.81s\nCreate floating ip address ---------------------------------------------- 5.34s\nCreate test server group ------------------------------------------------ 5.10s\nCreate ssh security group ----------------------------------------------- 4.93s\nCreate test network ----------------------------------------------------- 4.91s\nCreate test instances --------------------------------------------------- 4.78s\nAdd tag to instances ---------------------------------------------------- 4.61s\nAdd metadata to instances ----------------------------------------------- 4.39s\nAdd rule to ssh security group ------------------------------------------ 4.32s\nCreate test user -------------------------------------------------------- 4.19s\nCreate test-admin user -------------------------------------------------- 4.14s\nAdd rule to icmp security group ----------------------------------------- 4.09s\n+ server_list\n+ openstack --os-cloud test server list\n+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+\n| ID                                   | Name   | Status | Networks                             | Image                    | Flavor   |\n+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+\n| 9f84049c-743f-4895-8425-3c6d6b59650a | test-4 | ACTIVE | test=192.168.112.180, 192.168.200.63 | N/A (booted from volume) | SCS-1L-1 |\n| a86d3bc5-e53a-421d-bf49-90bac35835eb | test-3 | ACTIVE | test=192.168.112.146, 192.168.200.91 | N/A (booted from volume) | SCS-1L-1 |\n| 72fea215-b4a0-47d7-a5f6-1c164d423e5f | test   | ACTIVE | test=192.168.112.156, 192.168.200.3  | N/A (booted from volume) | SCS-1L-1 |\n| f05d3527-1651-466e-b2c0-b2e9034c1f08 | test-2 | ACTIVE | test=192.168.112.181, 192.168.200.17 | N/A (booted from volume) | SCS-1L-1 |\n| f6a0beab-2de6-40d7-8e1c-30e7df59b6ae | test-1 | ACTIVE | test=192.168.112.117, 192.168.200.89 | N/A (booted from volume) | SCS-1L-1 |\n+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+\n+ openstack --os-cloud test server show test\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-26T04:13:03.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.156, 192.168.200.3                                                                                                                                                                                                                                                                                                                                                 |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 3c4a154524be0044d9af6f6e2ac04bbe090637876489341638c46842                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 72fea215-b4a0-47d7-a5f6-1c164d423e5f                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='41d770c6-35d8-40d7-98a6-bf5de323a1d4'                                                                                                                                                                                                                                                                                                             |\n|                                     | delete_on_termination='False', id='c0b1f93f-9c66-4f28-98bf-725944fcdd59'                                                                                                                                                                                                                                                                                                            |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-1\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-26T04:13:05.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.117, 192.168.200.89                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 3c4a154524be0044d9af6f6e2ac04bbe090637876489341638c46842                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | f6a0beab-2de6-40d7-8e1c-30e7df59b6ae                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='8241a54e-f4c8-4656-ba7f-0cc497b28890'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-2\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-26T04:13:04.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.181, 192.168.200.17                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 674ca42258a0b4c8c8031bac2c2e2fbf6c22c7834e3bc95678cd74b8                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | f05d3527-1651-466e-b2c0-b2e9034c1f08                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='28433d04-ce59-4919-8a32-660a14b3fb94'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-3\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-26T04:13:06.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.146, 192.168.200.91                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-26T04:12:38Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 674ca42258a0b4c8c8031bac2c2e2fbf6c22c7834e3bc95678cd74b8                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | a86d3bc5-e53a-421d-bf49-90bac35835eb                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-26T04:13:37Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='7ff6c181-375d-43be-8b24-6e9effb98e3b'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-4\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-26T04:13:06.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.180, 192.168.200.63                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-26T04:12:40Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | ee0f370abe6f4901d5b81d816f3256dda7c87709752fbf58b38a05ee                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 9f84049c-743f-4895-8425-3c6d6b59650a                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-26T04:13:38Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='2af11777-ad16-40e3-aa06-7eb5dd4b823c'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.146\nPING 192.168.112.146 (192.168.112.146) 56(84) bytes of data.\n64 bytes from 192.168.112.146: icmp_seq=1 ttl=63 time=7.13 ms\n64 bytes from 192.168.112.146: icmp_seq=2 ttl=63 time=2.30 ms\n64 bytes from 192.168.112.146: icmp_seq=3 ttl=63 time=1.91 ms\n\n--- 192.168.112.146 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.905/3.779/7.133/2.376 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=8.63 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.41 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.84 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.836/4.292/8.629/3.075 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.181\nPING 192.168.112.181 (192.168.112.181) 56(84) bytes of data.\n64 bytes from 192.168.112.181: icmp_seq=1 ttl=63 time=5.35 ms\n64 bytes from 192.168.112.181: icmp_seq=2 ttl=63 time=2.32 ms\n64 bytes from 192.168.112.181: icmp_seq=3 ttl=63 time=2.04 ms\n\n--- 192.168.112.181 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.039/3.237/5.352/1.499 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.77 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.27 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.64 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.641/4.227/8.773/3.224 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.180\nPING 192.168.112.180 (192.168.112.180) 56(84) bytes of data.\n64 bytes from 192.168.112.180: icmp_seq=1 ttl=63 time=7.43 ms\n64 bytes from 192.168.112.180: icmp_seq=2 ttl=63 time=2.92 ms\n64 bytes from 192.168.112.180: icmp_seq=3 ttl=63 time=1.78 ms\n\n--- 192.168.112.180 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.782/4.043/7.431/2.439 ms\n+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# CHECK",
                                "",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "+ echo",
                                "+ echo '# CHECK'",
                                "+ echo",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver 9.5.0 5.0.0",
                                "",
                                "## Containers @ testbed-manager",
                                "",
                                "+ [[ 1 -eq -1 ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager ps",
                                "\u001b[32m2026-03-26 04:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m",
                                "CONTAINER ID   IMAGE                                                                            COMMAND                  CREATED             STATUS                 PORTS                                                                             NAMES",
                                "a6c0ff0cf266   registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_blackbox_exporter",
                                "9950518a1d95   registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20251130        \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_alertmanager",
                                "2d772ec2381d   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130            \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                                                                                             prometheus_cadvisor",
                                "3db09ddd1147   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130        \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                                                                                            prometheus_node_exporter",
                                "9171c7f54c01   registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20251130           \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                                                                                            prometheus_server",
                                "42b99ccdba78   registry.osism.tech/osism/cephclient:18.2.7                                      \"/usr/bin/dumb-init \u2026\"   About an hour ago   Up 59 minutes                                                                                            cephclient",
                                "14fa8b69c69f   registry.osism.tech/kolla/release/cron:3.0.20251130                              \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               cron",
                                "e7d025959a4d   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               kolla_toolbox",
                                "76b296741eb4   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                                                                                               fluentd",
                                "7e42ae5d1d9f   registry.osism.tech/osism/openstackclient:2024.2                                 \"/usr/bin/dumb-init \u2026\"   2 hours ago         Up 2 hours                                                                                               openstackclient",
                                "9c516885849a   phpmyadmin/phpmyadmin:5.2                                                        \"/docker-entrypoint.\u2026\"   2 hours ago         Up 2 hours (healthy)   80/tcp                                                                            phpmyadmin",
                                "218b44791923   registry.osism.tech/osism/homer:v25.10.1                                         \"/bin/sh /entrypoint\u2026\"   2 hours ago         Up 2 hours (healthy)   8080/tcp                                                                          homer",
                                "26eeba89e7d3   registry.osism.tech/osism/cgit:1.2.3                                             \"httpd-foreground\"       2 hours ago         Up 2 hours             80/tcp                                                                            cgit",
                                "01f240bfb490   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta                        \"entrypoint.sh -f /e\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:3128->3128/tcp                                                       squid",
                                "46bc210ef4f1   registry.osism.tech/osism/inventory-reconciler:0.20251130.0                      \"/sbin/tini -- /entr\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-inventory_reconciler-1",
                                "e5947914229b   registry.osism.tech/osism/ceph-ansible:0.20251130.0                              \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     ceph-ansible",
                                "80f4f4ccb433   registry.osism.tech/osism/osism-ansible:0.20251130.0                             \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osism-ansible",
                                "4cce80abddf5   registry.osism.tech/osism/osism-kubernetes:0.20251130.0                          \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osism-kubernetes",
                                "383c6d4109e1   registry.osism.tech/osism/kolla-ansible:0.20251130.0                             \"/entrypoint.sh osis\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     kolla-ansible",
                                "73c5df6619ee   registry.osism.tech/osism/ara-server:1.7.3                                       \"sh -c '/wait && /ru\u2026\"   2 hours ago         Up 2 hours (healthy)   8000/tcp                                                                          manager-ara-server-1",
                                "a1385669c48e   registry.osism.tech/osism/osism-frontend:0.20251130.1                            \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours             192.168.16.5:3000->3000/tcp                                                       osism-frontend",
                                "aa27c19ce640   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-listener-1",
                                "319e06da1161   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-flower-1",
                                "0084ba6d7452   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-beat-1",
                                "61ee0d48fd3c   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1",
                                "c5005444125f   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- sleep\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     osismclient",
                                "fa875137e67a   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine                         \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours (healthy)   6379/tcp                                                                          manager-redis-1",
                                "4bd574e40d14   registry.osism.tech/osism/osism:0.20251130.1                                     \"/sbin/tini -- osism\u2026\"   2 hours ago         Up 2 hours (healthy)                                                                                     manager-openstack-1",
                                "ee921f5b5733   registry.osism.tech/dockerhub/library/mariadb:11.8.4                             \"docker-entrypoint.s\u2026\"   2 hours ago         Up 2 hours (healthy)   3306/tcp                                                                          manager-mariadb-1",
                                "9ba1921448df   registry.osism.tech/dockerhub/library/traefik:v3.5.0                             \"/entrypoint.sh trae\u2026\"   2 hours ago         Up 2 hours (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik",
                                "",
                                "## Images @ testbed-manager",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager images",
                                "REPOSITORY                                                       TAG               IMAGE ID       CREATED         SIZE",
                                "registry.osism.tech/osism/openstackclient                        2024.2            1806a5037062   24 hours ago    239MB",
                                "registry.osism.tech/dockerhub/library/redis                      7.4.7-alpine      e08bd8d5a677   8 weeks ago     41.4MB",
                                "registry.osism.tech/osism/homer                                  v25.10.1          ea34b371c716   3 months ago    11.5MB",
                                "registry.osism.tech/osism/kolla-ansible                          0.20251130.0      0f140ec71e5f   3 months ago    608MB",
                                "registry.osism.tech/kolla/release/kolla-toolbox                  19.7.1.20251130   314d22193a72   3 months ago    669MB",
                                "registry.osism.tech/kolla/release/cron                           3.0.20251130      e1e0428a330f   3 months ago    265MB",
                                "registry.osism.tech/kolla/release/fluentd                        5.0.8.20251130    fb3c98fc8cae   3 months ago    578MB",
                                "registry.osism.tech/kolla/release/prometheus-blackbox-exporter   0.25.0.20251130   7bbb4f6f4831   3 months ago    308MB",
                                "registry.osism.tech/kolla/release/prometheus-cadvisor            0.49.2.20251130   591cbce746c1   3 months ago    357MB",
                                "registry.osism.tech/kolla/release/prometheus-alertmanager        0.28.0.20251130   ba994ea4acda   3 months ago    404MB",
                                "registry.osism.tech/kolla/release/prometheus-v2-server           2.55.1.20251130   56b43d5c716a   3 months ago    839MB",
                                "registry.osism.tech/kolla/release/prometheus-node-exporter       1.8.2.20251130    c1ab1d07f7ef   3 months ago    305MB",
                                "registry.osism.tech/osism/inventory-reconciler                   0.20251130.0      1bfc1dadeee1   3 months ago    330MB",
                                "registry.osism.tech/osism/osism-ansible                          0.20251130.0      42988b2d229c   3 months ago    613MB",
                                "registry.osism.tech/osism/ceph-ansible                           0.20251130.0      a212d8ca4a50   3 months ago    560MB",
                                "registry.osism.tech/osism/osism-kubernetes                       0.20251130.0      9beff03cb77b   3 months ago    1.23GB",
                                "registry.osism.tech/osism/osism                                  0.20251130.1      95213af683ec   3 months ago    383MB",
                                "registry.osism.tech/osism/osism-frontend                         0.20251130.1      2cb6e7609620   3 months ago    238MB",
                                "registry.osism.tech/dockerhub/library/mariadb                    11.8.4            70745dd8f1d0   4 months ago    334MB",
                                "phpmyadmin/phpmyadmin                                            5.2               e66b1f5a8c58   5 months ago    742MB",
                                "registry.osism.tech/osism/ara-server                             1.7.3             d1b687333f2f   7 months ago    275MB",
                                "registry.osism.tech/dockerhub/library/traefik                    v3.5.0            11cc59587f6a   8 months ago    226MB",
                                "registry.osism.tech/osism/cephclient                             18.2.7            ae977aa79826   10 months ago   453MB",
                                "registry.osism.tech/dockerhub/ubuntu/squid                       6.1-23.10_beta    34b6bbbcf74b   21 months ago   146MB",
                                "registry.osism.tech/osism/cgit                                   1.2.3             16e7285642b1   2 years ago     545MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver 9.5.0 5.0.0",
                                "",
                                "## Containers @ testbed-node-0",
                                "",
                                "+ [[ 1 -eq -1 ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 ps",
                                "CONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES",
                                "10e685d017e5   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor",
                                "1e8d61047d06   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api",
                                "db6b2a95a9bc   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   7 minutes ago       Up 7 minutes                           grafana",
                                "6542f78de3af   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter",
                                "467c18e96c3a   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor",
                                "137fab20c985   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_memcached_exporter",
                                "937717e9414d   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter",
                                "7eb22a81cd5e   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter",
                                "f3fd6a1d9bdd   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share",
                                "7f043dc613c0   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_scheduler",
                                "7d2ea26d8a03   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data",
                                "b1dd7de43631   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api",
                                "122590345af5   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier",
                                "c6af9a1009fc   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener",
                                "3a51fd5dcb26   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator",
                                "212c65b505ce   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api",
                                "428298fcfb6e   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central",
                                "6fff7b3529d3   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification",
                                "e0fc6d6cd546   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   21 minutes ago      Up 21 minutes (healthy)                octavia_worker",
                                "658463a5ab16   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping",
                                "1015fd0d4be5   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager",
                                "938dfb4b5809   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent",
                                "7b61d18683f1   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api",
                                "84d11367df1c   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker",
                                "f53fd21463f1   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns",
                                "6d3d7a997c3a   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer",
                                "7afbf1ab3e1d   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central",
                                "d021acb82cd1   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api",
                                "bbf56f666257   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9",
                                "02345889a066   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker",
                                "6690561e229b   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_keystone_listener",
                                "f3e2f038b0cd   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api",
                                "7eacd3eb081d   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup",
                                "33e9db06bbef   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_volume",
                                "501c5dc0acdd   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler",
                                "15d969d91992   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api",
                                "b158cf01e40b   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api",
                                "ee404b650bd7   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console",
                                "29a44929d5a6   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver",
                                "f16c1f62e400   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon",
                                "2415fb04f136   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy",
                                "7886f78fe202   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor",
                                "156eb6887714   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   44 minutes ago      Up 44 minutes (healthy)                nova_api",
                                "ab5d246d0721   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler",
                                "4ad302f098f7   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server",
                                "139e866a4b61   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   53 minutes ago      Up 53 minutes (healthy)                placement_api",
                                "1c6ed65052b6   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 55 minutes (healthy)                keystone",
                                "d393c5606814   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet",
                                "f820741ba15d   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh",
                                "198d679236a4   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-0",
                                "48ac09de19f5   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-0",
                                "c1b85917b265   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-0",
                                "28734fa9b704   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd",
                                "83ce56dde60c   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db",
                                "5d67cc327d74   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db",
                                "fdf645f7da16   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller",
                                "4e815f33dd1c   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd",
                                "5d315a0b0023   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db",
                                "196668baf1fa   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq",
                                "7d42071b582a   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb",
                                "b05b43a972e2   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel",
                                "e9731f9eb885   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis",
                                "9b8a2aa0aa07   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached",
                                "b1cb8702fe0a   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch_dashboards",
                                "05b2989d837f   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch",
                                "bb1665a8bd8c   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived",
                                "2f3893ca35e0   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql",
                                "9ccbd30640c7   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy",
                                "97d9ef633eb8   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron",
                                "cd160ef7c762   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox",
                                "dc42d1f63132   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd",
                                "",
                                "## Images @ testbed-node-0",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 images",
                                "REPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE",
                                "registry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB",
                                "registry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB",
                                "registry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB",
                                "registry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB",
                                "registry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB",
                                "registry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB",
                                "registry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB",
                                "registry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB",
                                "registry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB",
                                "registry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB",
                                "registry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB",
                                "registry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB",
                                "registry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB",
                                "registry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB",
                                "registry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB",
                                "registry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB",
                                "registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB",
                                "registry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB",
                                "registry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB",
                                "registry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB",
                                "registry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB",
                                "registry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB",
                                "registry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB",
                                "registry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB",
                                "registry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB",
                                "registry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB",
                                "registry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB",
                                "registry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB",
                                "registry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB",
                                "registry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB",
                                "registry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB",
                                "registry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB",
                                "registry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver 9.5.0 5.0.0",
                                "",
                                "## Containers @ testbed-node-1",
                                "",
                                "+ [[ 1 -eq -1 ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 ps",
                                "CONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES",
                                "5595f7aa7bde   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor",
                                "e70175a7f54f   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api",
                                "3947fc72dc2b   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   6 minutes ago       Up 6 minutes                           grafana",
                                "79d75ef2089b   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter",
                                "7bfe76f30932   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor",
                                "c528ae2db2f9   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_memcached_exporter",
                                "34200a94a9bb   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter",
                                "a619c7d8b861   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter",
                                "31eed0c039c6   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share",
                                "d7c831ca33ae   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_scheduler",
                                "ac2ea9870602   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data",
                                "a5e88825bb8f   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api",
                                "7e841d0eab1a   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier",
                                "281374c8a420   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener",
                                "c732d68cf8a0   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator",
                                "c34a9fd30b59   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api",
                                "8438fd6d7aab   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central",
                                "30fc5e345a33   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification",
                                "6eb7797a56dd   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_worker",
                                "0fa943cb70b5   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping",
                                "9ccee7b0750c   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager",
                                "ff30b926a413   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent",
                                "dbbf2a638521   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api",
                                "ab634d007474   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker",
                                "80b69c1f45f8   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns",
                                "0c654d6208e2   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer",
                                "c9a047b3172f   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central",
                                "c3785a6cb03e   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api",
                                "8f6b5ce65f81   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9",
                                "b5df48747341   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker",
                                "97c106d5b111   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_keystone_listener",
                                "b695ff5dbfeb   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api",
                                "3740fc0af171   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup",
                                "6e071ad5926e   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 31 minutes (healthy)                cinder_volume",
                                "e794b78209c8   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler",
                                "acd586036117   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api",
                                "e56028a9ebc8   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api",
                                "06f4dc20a799   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console",
                                "b30b39ae1f62   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver",
                                "12aa037e6068   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon",
                                "5905244ae562   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy",
                                "3f86776bbdd0   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor",
                                "39ee1db54373   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   45 minutes ago      Up 44 minutes (healthy)                nova_api",
                                "17d85da2dc3e   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler",
                                "8ecb795d537e   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server",
                                "2c6fcd0fb49d   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   53 minutes ago      Up 53 minutes (healthy)                placement_api",
                                "2f709020f008   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone",
                                "eded742115a7   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet",
                                "6c890b259434   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh",
                                "ef5500dbe92f   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-1",
                                "4f9927e71ae2   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-1",
                                "1fb5a820b9f6   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-1",
                                "7afd57013429   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd",
                                "eb26e25add2e   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db",
                                "4474d5901cb6   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db",
                                "f3366fda0db6   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller",
                                "8eea01b43193   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd",
                                "fd6d0d5f070c   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db",
                                "cbec62d736b0   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq",
                                "a2c9b8c73ba7   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb",
                                "de191e12b2ba   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel",
                                "f4bf64deb003   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis",
                                "04c02bb470b8   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached",
                                "c5a6b7eb3732   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             opensearch_dashboards",
                                "16973c91b145   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch",
                                "88a35f9e4664   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived",
                                "bf5b20e3ad39   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql",
                                "a55c8b3a0c88   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy",
                                "b85bba8a226d   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron",
                                "bfa74148a0c1   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox",
                                "0f65f0407b45   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd",
                                "",
                                "## Images @ testbed-node-1",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 images",
                                "REPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE",
                                "registry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB",
                                "registry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB",
                                "registry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB",
                                "registry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB",
                                "registry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB",
                                "registry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB",
                                "registry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB",
                                "registry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB",
                                "registry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB",
                                "registry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB",
                                "registry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB",
                                "registry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB",
                                "registry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB",
                                "registry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB",
                                "registry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB",
                                "registry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB",
                                "registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB",
                                "registry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB",
                                "registry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB",
                                "registry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB",
                                "registry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB",
                                "registry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB",
                                "registry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB",
                                "registry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB",
                                "registry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB",
                                "registry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB",
                                "registry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB",
                                "registry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB",
                                "registry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB",
                                "registry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB",
                                "registry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB",
                                "registry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB",
                                "registry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver 9.5.0 5.0.0",
                                "",
                                "## Containers @ testbed-node-2",
                                "",
                                "+ [[ 1 -eq -1 ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 ps",
                                "CONTAINER ID   IMAGE                                                                                COMMAND                  CREATED             STATUS                       PORTS     NAMES",
                                "cb5a2e0036ae   registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20251130                   \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_conductor",
                                "b1fc9f76bfe1   registry.osism.tech/kolla/release/magnum-api:19.0.1.20251130                         \"dumb-init --single-\u2026\"   3 minutes ago       Up 3 minutes (healthy)                 magnum_api",
                                "79bb3cfa27d7   registry.osism.tech/kolla/release/grafana:12.3.0.20251130                            \"dumb-init --single-\u2026\"   6 minutes ago       Up 6 minutes                           grafana",
                                "74f7548606e0   registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20251130   \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_elasticsearch_exporter",
                                "6ff2ddb90324   registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130                \"dumb-init --single-\u2026\"   9 minutes ago       Up 9 minutes                           prometheus_cadvisor",
                                "7541b568abd6   registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20251130      \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_memcached_exporter",
                                "c8e9e481f34b   registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20251130         \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_mysqld_exporter",
                                "91072cae9492   registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130            \"dumb-init --single-\u2026\"   10 minutes ago      Up 10 minutes                          prometheus_node_exporter",
                                "e9bb963e7218   registry.osism.tech/kolla/release/manila-share:19.1.1.20251130                       \"dumb-init --single-\u2026\"   14 minutes ago      Up 14 minutes (healthy)                manila_share",
                                "12a9083a7490   registry.osism.tech/kolla/release/manila-scheduler:19.1.1.20251130                   \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_scheduler",
                                "4f365a87ab5f   registry.osism.tech/kolla/release/manila-data:19.1.1.20251130                        \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_data",
                                "42b0d560762f   registry.osism.tech/kolla/release/manila-api:19.1.1.20251130                         \"dumb-init --single-\u2026\"   15 minutes ago      Up 15 minutes (healthy)                manila_api",
                                "4511ec114d0d   registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_notifier",
                                "39c62c61e597   registry.osism.tech/kolla/release/aodh-listener:19.0.0.20251130                      \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_listener",
                                "4eb2be969478   registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20251130                     \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_evaluator",
                                "3dd95017f326   registry.osism.tech/kolla/release/aodh-api:19.0.0.20251130                           \"dumb-init --single-\u2026\"   18 minutes ago      Up 18 minutes (healthy)                aodh_api",
                                "93d1c0dc0e86   registry.osism.tech/kolla/release/ceilometer-central:23.0.2.20251130                 \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes                          ceilometer_central",
                                "c225d72455b3   registry.osism.tech/kolla/release/ceilometer-notification:23.0.2.20251130            \"dumb-init --single-\u2026\"   20 minutes ago      Up 20 minutes (healthy)                ceilometer_notification",
                                "a01140bcf6bf   registry.osism.tech/kolla/release/octavia-worker:15.0.2.20251130                     \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_worker",
                                "0056447f6cdc   registry.osism.tech/kolla/release/octavia-housekeeping:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_housekeeping",
                                "23eb3d9e27c0   registry.osism.tech/kolla/release/octavia-health-manager:15.0.2.20251130             \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_health_manager",
                                "649fe0554c73   registry.osism.tech/kolla/release/octavia-driver-agent:15.0.2.20251130               \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes                          octavia_driver_agent",
                                "87e3f0068950   registry.osism.tech/kolla/release/octavia-api:15.0.2.20251130                        \"dumb-init --single-\u2026\"   22 minutes ago      Up 22 minutes (healthy)                octavia_api",
                                "dc2d6c4aae9a   registry.osism.tech/kolla/release/designate-worker:19.0.1.20251130                   \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_worker",
                                "cd5dacfed1b8   registry.osism.tech/kolla/release/designate-mdns:19.0.1.20251130                     \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_mdns",
                                "c0e8307f4672   registry.osism.tech/kolla/release/designate-producer:19.0.1.20251130                 \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_producer",
                                "68bd368adfdc   registry.osism.tech/kolla/release/designate-central:19.0.1.20251130                  \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_central",
                                "c51595c11699   registry.osism.tech/kolla/release/designate-api:19.0.1.20251130                      \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_api",
                                "8dffa9e9b29d   registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20251130            \"dumb-init --single-\u2026\"   27 minutes ago      Up 27 minutes (healthy)                designate_backend_bind9",
                                "6d0216ea6747   registry.osism.tech/kolla/release/barbican-worker:19.0.1.20251130                    \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes (healthy)                barbican_worker",
                                "2ad9f138b268   registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20251130         \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_keystone_listener",
                                "61e3accb5af7   registry.osism.tech/kolla/release/barbican-api:19.0.1.20251130                       \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes (healthy)                barbican_api",
                                "4c7ed3dcd5fc   registry.osism.tech/kolla/release/cinder-backup:25.3.1.20251130                      \"dumb-init --single-\u2026\"   31 minutes ago      Up 31 minutes (healthy)                cinder_backup",
                                "cf0ff34a99c9   registry.osism.tech/kolla/release/cinder-volume:25.3.1.20251130                      \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_volume",
                                "67e117d5e66e   registry.osism.tech/kolla/release/cinder-scheduler:25.3.1.20251130                   \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_scheduler",
                                "70fa2260ff57   registry.osism.tech/kolla/release/cinder-api:25.3.1.20251130                         \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes (healthy)                cinder_api",
                                "799f8b17e49d   registry.osism.tech/kolla/release/glance-api:29.0.1.20251130                         \"dumb-init --single-\u2026\"   34 minutes ago      Up 34 minutes (healthy)                glance_api",
                                "1230e722ebb3   registry.osism.tech/kolla/release/skyline-console:5.0.1.20251130                     \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_console",
                                "1b7da13de50d   registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20251130                   \"dumb-init --single-\u2026\"   37 minutes ago      Up 37 minutes (healthy)                skyline_apiserver",
                                "8142e29746fa   registry.osism.tech/kolla/release/horizon:25.1.2.20251130                            \"dumb-init --single-\u2026\"   39 minutes ago      Up 39 minutes (healthy)                horizon",
                                "13a1d439e229   registry.osism.tech/kolla/release/nova-novncproxy:30.2.1.20251130                    \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_novncproxy",
                                "3df27d3081d1   registry.osism.tech/kolla/release/nova-conductor:30.2.1.20251130                     \"dumb-init --single-\u2026\"   43 minutes ago      Up 43 minutes (healthy)                nova_conductor",
                                "02286aff1530   registry.osism.tech/kolla/release/nova-api:30.2.1.20251130                           \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_api",
                                "69a542a499a4   registry.osism.tech/kolla/release/nova-scheduler:30.2.1.20251130                     \"dumb-init --single-\u2026\"   45 minutes ago      Up 45 minutes (healthy)                nova_scheduler",
                                "04863e81a670   registry.osism.tech/kolla/release/neutron-server:25.2.2.20251130                     \"dumb-init --single-\u2026\"   50 minutes ago      Up 50 minutes (healthy)                neutron_server",
                                "943c7539f102   registry.osism.tech/kolla/release/placement-api:12.0.1.20251130                      \"dumb-init --single-\u2026\"   54 minutes ago      Up 54 minutes (healthy)                placement_api",
                                "1795602436d0   registry.osism.tech/kolla/release/keystone:26.0.1.20251130                           \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone",
                                "390c789c5ca0   registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20251130                    \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_fernet",
                                "32eece189277   registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20251130                       \"dumb-init --single-\u2026\"   56 minutes ago      Up 56 minutes (healthy)                keystone_ssh",
                                "4373766b4740   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mgr -\u2026\"   58 minutes ago      Up 58 minutes                          ceph-mgr-testbed-node-2",
                                "c7fe63b73697   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-crash\"    About an hour ago   Up About an hour                       ceph-crash-testbed-node-2",
                                "2a382ea60872   registry.osism.tech/osism/ceph-daemon:18.2.7                                         \"/usr/bin/ceph-mon -\u2026\"   About an hour ago   Up About an hour                       ceph-mon-testbed-node-2",
                                "09e52c0af1ac   registry.osism.tech/kolla/release/ovn-northd:24.9.3.20251130                         \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_northd",
                                "5cd3e7218a58   registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_sb_db",
                                "6a6022c2623e   registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.3.20251130                   \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_nb_db",
                                "0e57fca16035   registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour                       ovn_controller",
                                "b5135bee0b88   registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130                \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_vswitchd",
                                "2e496286f4d6   registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130               \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             openvswitch_db",
                                "774e79736876   registry.osism.tech/kolla/release/rabbitmq:3.13.7.20251130                           \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             rabbitmq",
                                "a33e8c9a11ce   registry.osism.tech/kolla/release/mariadb-server:10.11.15.20251130                   \"dumb-init -- kolla_\u2026\"   About an hour ago   Up About an hour (healthy)             mariadb",
                                "dc6520f0098a   registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20251130                     \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis_sentinel",
                                "620d722dcf31   registry.osism.tech/kolla/release/redis:7.0.15.20251130                              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             redis",
                                "a26c32c37c05   registry.osism.tech/kolla/release/memcached:1.6.24.20251130                          \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             memcached",
                                "5df26a02d819   registry.osism.tech/kolla/release/opensearch-dashboards:2.19.4.20251130              \"dumb-init --single-\u2026\"   About an hour ago   Up About an hour (healthy)             opensearch_dashboards",
                                "a0ce1c2c8df2   registry.osism.tech/kolla/release/opensearch:2.19.4.20251130                         \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   opensearch",
                                "35762f95d946   registry.osism.tech/kolla/release/keepalived:2.2.8.20251130                          \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             keepalived",
                                "4a8d2c92eea7   registry.osism.tech/kolla/release/proxysql:3.0.3.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   proxysql",
                                "6a2161fe9af1   registry.osism.tech/kolla/release/haproxy:2.8.15.20251130                            \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours (healthy)                   haproxy",
                                "c05e2866aaa1   registry.osism.tech/kolla/release/cron:3.0.20251130                                  \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             cron",
                                "d6f4b784a67c   registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130                      \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             kolla_toolbox",
                                "7742d32707c2   registry.osism.tech/kolla/release/fluentd:5.0.8.20251130                             \"dumb-init --single-\u2026\"   2 hours ago         Up 2 hours                             fluentd",
                                "",
                                "## Images @ testbed-node-2",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 images",
                                "REPOSITORY                                                            TAG                 IMAGE ID       CREATED         SIZE",
                                "registry.osism.tech/kolla/release/rabbitmq                            3.13.7.20251130     618df24dfbf4   3 months ago    322MB",
                                "registry.osism.tech/kolla/release/memcached                           1.6.24.20251130     8a9865997707   3 months ago    266MB",
                                "registry.osism.tech/kolla/release/opensearch                          2.19.4.20251130     dc62f23331d2   3 months ago    1.56GB",
                                "registry.osism.tech/kolla/release/keepalived                          2.2.8.20251130      94862d07fc5a   3 months ago    276MB",
                                "registry.osism.tech/kolla/release/opensearch-dashboards               2.19.4.20251130     3b3613dd9b1a   3 months ago    1.53GB",
                                "registry.osism.tech/kolla/release/kolla-toolbox                       19.7.1.20251130     314d22193a72   3 months ago    669MB",
                                "registry.osism.tech/kolla/release/cron                                3.0.20251130        e1e0428a330f   3 months ago    265MB",
                                "registry.osism.tech/kolla/release/grafana                             12.3.0.20251130     6eb3b7b1dbf2   3 months ago    1.02GB",
                                "registry.osism.tech/kolla/release/proxysql                            3.0.3.20251130      2c7177938c0e   3 months ago    412MB",
                                "registry.osism.tech/kolla/release/haproxy                             2.8.15.20251130     6d4c583df983   3 months ago    274MB",
                                "registry.osism.tech/kolla/release/fluentd                             5.0.8.20251130      fb3c98fc8cae   3 months ago    578MB",
                                "registry.osism.tech/kolla/release/redis                               7.0.15.20251130     5548a8ce5b5c   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/redis-sentinel                      7.0.15.20251130     62d0b016058f   3 months ago    273MB",
                                "registry.osism.tech/kolla/release/mariadb-server                      10.11.15.20251130   77db67eebcc3   3 months ago    452MB",
                                "registry.osism.tech/kolla/release/horizon                             25.1.2.20251130     d7257ed845e9   3 months ago    1.15GB",
                                "registry.osism.tech/kolla/release/prometheus-mysqld-exporter          0.16.0.20251130     aedc672fb472   3 months ago    301MB",
                                "registry.osism.tech/kolla/release/prometheus-memcached-exporter       0.15.0.20251130     7b077076926d   3 months ago    298MB",
                                "registry.osism.tech/kolla/release/prometheus-cadvisor                 0.49.2.20251130     591cbce746c1   3 months ago    357MB",
                                "registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter   1.8.0.20251130      bcaaf5d64345   3 months ago    292MB",
                                "registry.osism.tech/kolla/release/prometheus-node-exporter            1.8.2.20251130      c1ab1d07f7ef   3 months ago    305MB",
                                "registry.osism.tech/kolla/release/openvswitch-db-server               3.4.3.20251130      3e6f3fe8823c   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/placement-api                       12.0.1.20251130     20317ff6dfb9   3 months ago    975MB",
                                "registry.osism.tech/kolla/release/openvswitch-vswitchd                3.4.3.20251130      ad8bb4636454   3 months ago    279MB",
                                "registry.osism.tech/kolla/release/nova-novncproxy                     30.2.1.20251130     99323056afa4   3 months ago    1.37GB",
                                "registry.osism.tech/kolla/release/nova-conductor                      30.2.1.20251130     92609e648215   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-api                            30.2.1.20251130     2d78e7fdfb9a   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/nova-scheduler                      30.2.1.20251130     4c3c59730530   3 months ago    1.21GB",
                                "registry.osism.tech/kolla/release/ceilometer-central                  23.0.2.20251130     37cb6975d4a5   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/ceilometer-notification             23.0.2.20251130     bb2927b293dc   3 months ago    976MB",
                                "registry.osism.tech/kolla/release/magnum-api                          19.0.1.20251130     a85fdbb4bbba   3 months ago    1.13GB",
                                "registry.osism.tech/kolla/release/magnum-conductor                    19.0.1.20251130     a98ee1099aad   3 months ago    1.24GB",
                                "registry.osism.tech/kolla/release/manila-share                        19.1.1.20251130     df44f491f2c1   3 months ago    1.22GB",
                                "registry.osism.tech/kolla/release/manila-data                         19.1.1.20251130     cd8b74c8a47a   3 months ago    1.06GB",
                                "registry.osism.tech/kolla/release/manila-api                          19.1.1.20251130     654f9bd3c940   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/manila-scheduler                    19.1.1.20251130     e0864fa03a78   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/aodh-listener                       19.0.0.20251130     1e68c23a9d38   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-evaluator                      19.0.0.20251130     1726a7592f93   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-notifier                       19.0.0.20251130     abbd6e9f87e2   3 months ago    974MB",
                                "registry.osism.tech/kolla/release/aodh-api                            19.0.0.20251130     82a64f1d056d   3 months ago    973MB",
                                "registry.osism.tech/kolla/release/barbican-worker                     19.0.1.20251130     2cef5d51872b   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-keystone-listener          19.0.1.20251130     bfcd8631a126   3 months ago    991MB",
                                "registry.osism.tech/kolla/release/barbican-api                        19.0.1.20251130     9195ddc3e4c5   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/keystone                            26.0.1.20251130     6c1543e94c06   3 months ago    1.09GB",
                                "registry.osism.tech/kolla/release/keystone-fernet                     26.0.1.20251130     36669c355898   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/keystone-ssh                        26.0.1.20251130     e002cffc8eb8   3 months ago    1.04GB",
                                "registry.osism.tech/kolla/release/octavia-health-manager              15.0.2.20251130     059dc6d4a159   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-housekeeping                15.0.2.20251130     c9059accdc4a   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-api                         15.0.2.20251130     9375641bed7a   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/octavia-worker                      15.0.2.20251130     708f50e37fa7   3 months ago    1.03GB",
                                "registry.osism.tech/kolla/release/octavia-driver-agent                15.0.2.20251130     045f928baedc   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/neutron-server                      25.2.2.20251130     fa71fe0a109e   3 months ago    1.16GB",
                                "registry.osism.tech/kolla/release/glance-api                          29.0.1.20251130     b1fcfbc49057   3 months ago    1.1GB",
                                "registry.osism.tech/kolla/release/designate-central                   19.0.1.20251130     00b6af03994a   3 months ago    983MB",
                                "registry.osism.tech/kolla/release/designate-worker                    19.0.1.20251130     18bc80370e46   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-producer                  19.0.1.20251130     eac4506bf51f   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-api                       19.0.1.20251130     ad5d5cd1392a   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/designate-backend-bind9             19.0.1.20251130     4e19a1dc9c8a   3 months ago    989MB",
                                "registry.osism.tech/kolla/release/designate-mdns                      19.0.1.20251130     4ad9e0017d6e   3 months ago    984MB",
                                "registry.osism.tech/kolla/release/skyline-console                     5.0.1.20251130      20430a0acd38   3 months ago    1.05GB",
                                "registry.osism.tech/kolla/release/skyline-apiserver                   5.0.1.20251130      20bbe1600b66   3 months ago    990MB",
                                "registry.osism.tech/kolla/release/cinder-volume                       25.3.1.20251130     ab7ee3c06214   3 months ago    1.72GB",
                                "registry.osism.tech/kolla/release/cinder-scheduler                    25.3.1.20251130     47d31cd2c25d   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/cinder-backup                       25.3.1.20251130     c09074b62f18   3 months ago    1.41GB",
                                "registry.osism.tech/kolla/release/cinder-api                          25.3.1.20251130     ceaaac81e8af   3 months ago    1.4GB",
                                "registry.osism.tech/kolla/release/ovn-nb-db-server                    24.9.3.20251130     e52b6499881a   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-controller                      24.9.3.20251130     fcd09e53d925   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-sb-db-server                    24.9.3.20251130     2fcefdb5b030   3 months ago    840MB",
                                "registry.osism.tech/kolla/release/ovn-northd                          24.9.3.20251130     948e5d22de86   3 months ago    840MB",
                                "registry.osism.tech/osism/ceph-daemon                                 18.2.7              5f92363b1f93   10 months ago   1.27GB",
                                "+ sh -c /opt/configuration/scripts/check-services.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# Ceph status",
                                "",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "+ echo",
                                "+ echo '# Ceph status'",
                                "+ echo",
                                "+ ceph -s",
                                "  cluster:",
                                "    id:     11111111-1111-1111-1111-111111111111",
                                "    health: HEALTH_OK",
                                " ",
                                "  services:",
                                "    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 71m)",
                                "    mgr: testbed-node-1(active, since 58m), standbys: testbed-node-0, testbed-node-2",
                                "    mds: 1/1 daemons up, 2 standby",
                                "    osd: 6 osds: 6 up (since 67m), 6 in (since 68m)",
                                "    rgw: 3 daemons active (3 hosts, 1 zones)",
                                " ",
                                "  data:",
                                "    volumes: 1/1 healthy",
                                "    pools:   14 pools, 401 pgs",
                                "    objects: 556 objects, 2.2 GiB",
                                "    usage:   7.0 GiB used, 113 GiB / 120 GiB avail",
                                "    pgs:     401 active+clean",
                                " ",
                                "",
                                "# Ceph versions",
                                "",
                                "+ echo",
                                "+ echo '# Ceph versions'",
                                "+ echo",
                                "+ ceph versions",
                                "{",
                                "    \"mon\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"mgr\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"osd\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6",
                                "    },",
                                "    \"mds\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"rgw\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"overall\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18",
                                "    }",
                                "}",
                                "",
                                "# Ceph OSD tree",
                                "",
                                "+ echo",
                                "+ echo '# Ceph OSD tree'",
                                "+ echo",
                                "+ ceph osd df tree",
                                "ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              ",
                                "-1         0.11691         -  120 GiB  7.0 GiB  6.7 GiB    6 KiB  369 MiB  113 GiB  5.87  1.00    -          root default           ",
                                "-3         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-3",
                                " 0    hdd  0.01949   1.00000   20 GiB  840 MiB  779 MiB    1 KiB   62 MiB   19 GiB  4.11  0.70  189      up          osd.0          ",
                                " 3    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.5 GiB    1 KiB   62 MiB   18 GiB  7.63  1.30  201      up          osd.3          ",
                                "-5         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-4",
                                " 2    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.4 GiB    1 KiB   62 MiB   19 GiB  7.20  1.23  206      up          osd.2          ",
                                " 5    hdd  0.01949   1.00000   20 GiB  928 MiB  867 MiB    1 KiB   62 MiB   19 GiB  4.54  0.77  186      up          osd.5          ",
                                "-7         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  123 MiB   38 GiB  5.87  1.00    -              host testbed-node-5",
                                " 1    hdd  0.01949   1.00000   20 GiB  1.2 GiB  1.1 GiB    1 KiB   62 MiB   19 GiB  5.95  1.01  192      up          osd.1          ",
                                " 4    hdd  0.01949   1.00000   20 GiB  1.2 GiB  1.1 GiB    1 KiB   62 MiB   19 GiB  5.79  0.99  196      up          osd.4          ",
                                "                       TOTAL  120 GiB  7.0 GiB  6.7 GiB  9.3 KiB  369 MiB  113 GiB  5.87                                            ",
                                "MIN/MAX VAR: 0.70/1.30  STDDEV: 1.27",
                                "",
                                "# Ceph monitor status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph monitor status'",
                                "+ echo",
                                "+ ceph mon stat",
                                "e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2",
                                "",
                                "# Ceph quorum status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph quorum status'",
                                "+ echo",
                                "+ ceph quorum_status",
                                "+ jq",
                                "{",
                                "  \"election_epoch\": 8,",
                                "  \"quorum\": [",
                                "    0,",
                                "    1,",
                                "    2",
                                "  ],",
                                "  \"quorum_names\": [",
                                "    \"testbed-node-0\",",
                                "    \"testbed-node-1\",",
                                "    \"testbed-node-2\"",
                                "  ],",
                                "  \"quorum_leader_name\": \"testbed-node-0\",",
                                "  \"quorum_age\": 4267,",
                                "  \"features\": {",
                                "    \"quorum_con\": \"4540138322906710015\",",
                                "    \"quorum_mon\": [",
                                "      \"kraken\",",
                                "      \"luminous\",",
                                "      \"mimic\",",
                                "      \"osdmap-prune\",",
                                "      \"nautilus\",",
                                "      \"octopus\",",
                                "      \"pacific\",",
                                "      \"elector-pinging\",",
                                "      \"quincy\",",
                                "      \"reef\"",
                                "    ]",
                                "  },",
                                "  \"monmap\": {",
                                "    \"epoch\": 1,",
                                "    \"fsid\": \"11111111-1111-1111-1111-111111111111\",",
                                "    \"modified\": \"2026-03-26T02:53:06.287135Z\",",
                                "    \"created\": \"2026-03-26T02:53:06.287135Z\",",
                                "    \"min_mon_release\": 18,",
                                "    \"min_mon_release_name\": \"reef\",",
                                "    \"election_strategy\": 1,",
                                "    \"disallowed_leaders: \": \"\",",
                                "    \"stretch_mode\": false,",
                                "    \"tiebreaker_mon\": \"\",",
                                "    \"removed_ranks: \": \"\",",
                                "    \"features\": {",
                                "      \"persistent\": [",
                                "        \"kraken\",",
                                "        \"luminous\",",
                                "        \"mimic\",",
                                "        \"osdmap-prune\",",
                                "        \"nautilus\",",
                                "        \"octopus\",",
                                "        \"pacific\",",
                                "        \"elector-pinging\",",
                                "        \"quincy\",",
                                "        \"reef\"",
                                "      ],",
                                "      \"optional\": []",
                                "    },",
                                "    \"mons\": [",
                                "      {",
                                "        \"rank\": 0,",
                                "        \"name\": \"testbed-node-0\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.10:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.10:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.10:6789/0\",",
                                "        \"public_addr\": \"192.168.16.10:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 1,",
                                "        \"name\": \"testbed-node-1\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.11:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.11:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.11:6789/0\",",
                                "        \"public_addr\": \"192.168.16.11:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 2,",
                                "        \"name\": \"testbed-node-2\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.12:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.12:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.12:6789/0\",",
                                "        \"public_addr\": \"192.168.16.12:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      }",
                                "    ]",
                                "  }",
                                "}",
                                "",
                                "# Ceph free space status",
                                "+ echo",
                                "+ echo '# Ceph free space status'",
                                "",
                                "+ echo",
                                "+ ceph df",
                                "--- RAW STORAGE ---",
                                "CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED",
                                "hdd    120 GiB  113 GiB  7.0 GiB   7.0 GiB       5.87",
                                "TOTAL  120 GiB  113 GiB  7.0 GiB   7.0 GiB       5.87",
                                " ",
                                "--- POOLS ---",
                                "POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL",
                                ".mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB",
                                "cephfs_data                 2   32      0 B        0      0 B      0     35 GiB",
                                "cephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB",
                                "default.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.control         6   32      0 B        8      0 B      0     35 GiB",
                                "default.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB",
                                "default.rgw.meta            8   32      0 B        0      0 B      0     35 GiB",
                                ".rgw.root                   9   32  3.9 KiB        8   64 KiB      0     52 GiB",
                                "backups                    10   32     19 B        2   12 KiB      0     35 GiB",
                                "volumes                    11   32     19 B        2   12 KiB      0     35 GiB",
                                "images                     12   32  2.2 GiB      299  6.7 GiB   5.98     35 GiB",
                                "metrics                    13   32     19 B        2   12 KiB      0     35 GiB",
                                "vms                        14   32     19 B        2   12 KiB      0     35 GiB",
                                "++ semver 9.5.0 5.0.0",
                                "+ [[ 1 -eq -1 ]]",
                                "+ [[ ! -e /etc/redhat-release ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-03-26 04:04:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8207bc9-b917-4707-81f1-417025aac495 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:04:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a8207bc9-b917-4707-81f1-417025aac495 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  04:04:38 +0000 (0:00:00.296)       0:00:00.296 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Thursday 26 March 2026  04:04:39 +0000 (0:00:01.129)       0:00:01.425 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  04:04:41 +0000 (0:00:01.360)       0:00:02.785 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Thursday 26 March 2026  04:04:46 +0000 (0:00:05.714)       0:00:08.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:04:47 +0000 (0:00:00.589)       0:00:09.089 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.71s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.36s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.13s",
                                "Gather facts for all hosts ---------------------------------------------- 0.59s",
                                "+ osism validate ceph-mons",
                                "",
                                "PLAY [Ceph validate mons] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Thursday 26 March 2026  04:05:04 +0000 (0:00:00.452)       0:00:00.452 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Thursday 26 March 2026  04:05:05 +0000 (0:00:00.853)       0:00:01.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Thursday 26 March 2026  04:05:06 +0000 (0:00:00.812)       0:00:02.118 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Thursday 26 March 2026  04:05:06 +0000 (0:00:00.115)       0:00:02.234 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Thursday 26 March 2026  04:05:06 +0000 (0:00:00.259)       0:00:02.494 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Thursday 26 March 2026  04:05:07 +0000 (0:00:01.012)       0:00:03.506 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Thursday 26 March 2026  04:05:08 +0000 (0:00:00.267)       0:00:03.774 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:05:08 +0000 (0:00:00.407)       0:00:04.181 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mon is not running] ********************",
                                "Thursday 26 March 2026  04:05:08 +0000 (0:00:00.299)       0:00:04.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mon is running] ************************",
                                "Thursday 26 March 2026  04:05:09 +0000 (0:00:00.289)       0:00:04.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:05:09 +0000 (0:00:00.426)       0:00:05.196 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:05:09 +0000 (0:00:00.225)       0:00:05.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:05:09 +0000 (0:00:00.233)       0:00:05.655 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.261)       0:00:05.916 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.068)       0:00:05.985 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.071)       0:00:06.056 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.071)       0:00:06.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.225)       0:00:06.353 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare quorum test vars] ************************************************",
                                "Thursday 26 March 2026  04:05:10 +0000 (0:00:00.234)       0:00:06.588 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get monmap info from one mon container] **********************************",
                                "Thursday 26 March 2026  04:05:11 +0000 (0:00:00.115)       0:00:06.703 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set quorum test data] ****************************************************",
                                "Thursday 26 March 2026  04:05:12 +0000 (0:00:01.639)       0:00:08.343 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail quorum test if not all monitors are in quorum] **********************",
                                "Thursday 26 March 2026  04:05:13 +0000 (0:00:00.528)       0:00:08.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass quorum test if all monitors are in quorum] **************************",
                                "Thursday 26 March 2026  04:05:13 +0000 (0:00:00.133)       0:00:09.004 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set fsid test vars] ******************************************************",
                                "Thursday 26 March 2026  04:05:13 +0000 (0:00:00.333)       0:00:09.338 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail Cluster FSID test if FSID does not match configuration] *************",
                                "Thursday 26 March 2026  04:05:13 +0000 (0:00:00.320)       0:00:09.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass Cluster FSID test if it matches configuration] **********************",
                                "Thursday 26 March 2026  04:05:14 +0000 (0:00:00.128)       0:00:09.787 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare status test vars] ************************************************",
                                "Thursday 26 March 2026  04:05:14 +0000 (0:00:00.116)       0:00:09.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather status data] ******************************************************",
                                "Thursday 26 March 2026  04:05:14 +0000 (0:00:00.168)       0:00:10.073 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set health test data] ****************************************************",
                                "Thursday 26 March 2026  04:05:15 +0000 (0:00:01.416)       0:00:11.490 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable] *************************",
                                "Thursday 26 March 2026  04:05:16 +0000 (0:00:00.313)       0:00:11.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if health is acceptable] *****************************",
                                "Thursday 26 March 2026  04:05:16 +0000 (0:00:00.143)       0:00:11.947 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable (strict)] ****************",
                                "Thursday 26 March 2026  04:05:16 +0000 (0:00:00.165)       0:00:12.112 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if status is OK (strict)] ****************************",
                                "Thursday 26 March 2026  04:05:16 +0000 (0:00:00.142)       0:00:12.255 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Thursday 26 March 2026  04:05:16 +0000 (0:00:00.384)       0:00:12.640 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Thursday 26 March 2026  04:05:17 +0000 (0:00:00.310)       0:00:12.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:05:17 +0000 (0:00:00.281)       0:00:13.231 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:05:19 +0000 (0:00:02.146)       0:00:15.378 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:05:19 +0000 (0:00:00.278)       0:00:15.657 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:20 +0000 (0:00:00.270)       0:00:15.927 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:20 +0000 (0:00:00.079)       0:00:16.007 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:20 +0000 (0:00:00.100)       0:00:16.107 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Thursday 26 March 2026  04:05:20 +0000 (0:00:00.084)       0:00:16.191 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:05:22 +0000 (0:00:01.618)       0:00:17.809 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-26T04:05:05+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:05:23 +0000 (0:00:00.917)       0:00:18.727 ******** ",
                                "=============================================================================== ",
                                "Aggregate test results step one ----------------------------------------- 2.15s",
                                "Get monmap info from one mon container ---------------------------------- 1.64s",
                                "Write report file ------------------------------------------------------- 1.62s",
                                "Gather status data ------------------------------------------------------ 1.42s",
                                "Get container info ------------------------------------------------------ 1.01s",
                                "Print report file information ------------------------------------------- 0.92s",
                                "Get timestamp for report file ------------------------------------------- 0.85s",
                                "Create report output directory ------------------------------------------ 0.81s",
                                "Set quorum test data ---------------------------------------------------- 0.53s",
                                "Set test result to passed if ceph-mon is running ------------------------ 0.43s",
                                "Set test result to passed if container is existing ---------------------- 0.41s",
                                "Pass cluster-health if status is OK (strict) ---------------------------- 0.38s",
                                "Pass quorum test if all monitors are in quorum -------------------------- 0.33s",
                                "Set fsid test vars ------------------------------------------------------ 0.32s",
                                "Set health test data ---------------------------------------------------- 0.31s",
                                "Set validation result to passed if no test failed ----------------------- 0.31s",
                                "Prepare test data ------------------------------------------------------- 0.30s",
                                "Set test result to failed if ceph-mon is not running -------------------- 0.29s",
                                "Set validation result to failed if a test failed ------------------------ 0.28s",
                                "Aggregate test results step two ----------------------------------------- 0.28s",
                                "+ osism validate ceph-mgrs",
                                "",
                                "PLAY [Ceph validate mgrs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Thursday 26 March 2026  04:05:40 +0000 (0:00:00.468)       0:00:00.468 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Thursday 26 March 2026  04:05:41 +0000 (0:00:00.832)       0:00:01.301 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Thursday 26 March 2026  04:05:42 +0000 (0:00:00.980)       0:00:02.281 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Thursday 26 March 2026  04:05:42 +0000 (0:00:00.146)       0:00:02.427 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Thursday 26 March 2026  04:05:42 +0000 (0:00:00.311)       0:00:02.739 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Thursday 26 March 2026  04:05:43 +0000 (0:00:01.008)       0:00:03.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Thursday 26 March 2026  04:05:44 +0000 (0:00:00.315)       0:00:04.064 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:05:44 +0000 (0:00:00.539)       0:00:04.603 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mgr is not running] ********************",
                                "Thursday 26 March 2026  04:05:44 +0000 (0:00:00.340)       0:00:04.943 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mgr is running] ************************",
                                "Thursday 26 March 2026  04:05:45 +0000 (0:00:00.299)       0:00:05.242 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:05:45 +0000 (0:00:00.544)       0:00:05.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.247)       0:00:06.035 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.249)       0:00:06.285 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.246)       0:00:06.532 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.070)       0:00:06.602 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.071)       0:00:06.674 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.075)       0:00:06.750 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Thursday 26 March 2026  04:05:46 +0000 (0:00:00.248)       0:00:06.999 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Define mgr module test vars] *********************************************",
                                "Thursday 26 March 2026  04:05:47 +0000 (0:00:00.245)       0:00:07.245 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather list of mgr modules] **********************************************",
                                "Thursday 26 March 2026  04:05:47 +0000 (0:00:00.118)       0:00:07.363 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Parse mgr module list from json] *****************************************",
                                "Thursday 26 March 2026  04:05:49 +0000 (0:00:02.082)       0:00:09.446 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Extract list of enabled mgr modules] *************************************",
                                "Thursday 26 March 2026  04:05:49 +0000 (0:00:00.462)       0:00:09.908 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail test if mgr modules are disabled that should be enabled] ************",
                                "Thursday 26 March 2026  04:05:50 +0000 (0:00:00.323)       0:00:10.231 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass test if required mgr modules are enabled] ***************************",
                                "Thursday 26 March 2026  04:05:50 +0000 (0:00:00.194)       0:00:10.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Thursday 26 March 2026  04:05:50 +0000 (0:00:00.143)       0:00:10.569 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Thursday 26 March 2026  04:05:50 +0000 (0:00:00.263)       0:00:10.832 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:05:51 +0000 (0:00:00.254)       0:00:11.087 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:05:52 +0000 (0:00:01.280)       0:00:12.368 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:05:52 +0000 (0:00:00.256)       0:00:12.625 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:52 +0000 (0:00:00.254)       0:00:12.879 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:52 +0000 (0:00:00.071)       0:00:12.951 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:05:53 +0000 (0:00:00.091)       0:00:13.042 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Thursday 26 March 2026  04:05:53 +0000 (0:00:00.270)       0:00:13.313 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:05:54 +0000 (0:00:01.330)       0:00:14.644 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-26T04:05:41+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:05:55 +0000 (0:00:00.405)       0:00:15.049 ******** ",
                                "=============================================================================== ",
                                "Gather list of mgr modules ---------------------------------------------- 2.08s",
                                "Write report file ------------------------------------------------------- 1.33s",
                                "Aggregate test results step one ----------------------------------------- 1.28s",
                                "Get container info ------------------------------------------------------ 1.01s",
                                "Create report output directory ------------------------------------------ 0.98s",
                                "Get timestamp for report file ------------------------------------------- 0.83s",
                                "Set test result to passed if ceph-mgr is running ------------------------ 0.54s",
                                "Set test result to passed if container is existing ---------------------- 0.54s",
                                "Parse mgr module list from json ----------------------------------------- 0.46s",
                                "Flush handlers ---------------------------------------------------------- 0.43s",
                                "Print report file information ------------------------------------------- 0.41s",
                                "Prepare test data ------------------------------------------------------- 0.34s",
                                "Extract list of enabled mgr modules ------------------------------------- 0.32s",
                                "Set test result to failed if container is missing ----------------------- 0.32s",
                                "Prepare test data for container existance test -------------------------- 0.31s",
                                "Set test result to failed if ceph-mgr is not running -------------------- 0.30s",
                                "Set validation result to passed if no test failed ----------------------- 0.26s",
                                "Aggregate test results step two ----------------------------------------- 0.26s",
                                "Set validation result to failed if a test failed ------------------------ 0.25s",
                                "Aggregate test results step three --------------------------------------- 0.25s",
                                "+ osism validate ceph-osds",
                                "",
                                "PLAY [Ceph validate OSDs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Thursday 26 March 2026  04:06:12 +0000 (0:00:00.441)       0:00:00.441 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Thursday 26 March 2026  04:06:13 +0000 (0:00:00.872)       0:00:01.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Thursday 26 March 2026  04:06:14 +0000 (0:00:00.554)       0:00:01.869 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Thursday 26 March 2026  04:06:14 +0000 (0:00:00.789)       0:00:02.658 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Thursday 26 March 2026  04:06:14 +0000 (0:00:00.133)       0:00:02.792 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Thursday 26 March 2026  04:06:15 +0000 (0:00:00.123)       0:00:02.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Thursday 26 March 2026  04:06:15 +0000 (0:00:00.312)       0:00:03.228 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Thursday 26 March 2026  04:06:15 +0000 (0:00:00.161)       0:00:03.390 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate total number of OSDs in cluster] *******************************",
                                "Thursday 26 March 2026  04:06:15 +0000 (0:00:00.315)       0:00:03.705 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:06:16 +0000 (0:00:00.784)       0:00:04.490 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers on host] *********************************",
                                "Thursday 26 March 2026  04:06:16 +0000 (0:00:00.292)       0:00:04.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '6a47ae157c609fdba1c3cb78cfabe35faae1300906bbf5c42bf9812cf5ae3fc3', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '19873f576682f61766adcb2a0a3bd7e7d79b6aaf7f359db744e235b802241d18', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '64dc82767cbb91bfba3d4fbe0f910f1c1b95d0856e1921a113524e50b04683ae', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '905865d8832bd6a48e47e77f25a307fabc6f4262f500433c693482e257e98351', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0a2ea494cefa3bc62f3a02d275955bf06259547cc9c9e7d9a6af1873fa1e568c', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'eedb79789b5f6a192a1b22b4016b1979f5d3369e8b3a5882bd980f73e378d679', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '99dc660051ff8aa712dc0eb018d8040e264e7abe5fee8210dbf2a9b2887da177', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b7b3e024c8e1cc2e99343381e2b424833301da2438f00dc4ab4f87d51ae0ce3d', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '261fb7fe11f046be78125f2f61829982321ff86ca846ffb773a8692a92cdc390', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a3e70f4017c0ded81315b1398083acb671abe365698c575bfc6fc4c9e4d55223', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'eacf9bcf824e7e1d8ff361b32cadee04f8af28293f8b9c40a61bc7a939d2150b', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': 'dd1dcbbb5855901401655533a58e1a34834b99fe4c43bf30a9fcba2126bd3e51', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': '30b29e57211792d4ff597f247627c18d4ebc0c7ea1e4f1feaa4a30caa4ab2f31', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ce801b2b1273df80d38b32bfc814e9fae0bff5a3f34069b7cd0581325b281bfc', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'c319b5f6aba28f04e9d72db4d66c7ea078f78b2519ee091016ef2335532f1d29', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a8c7c868a3e502d499713dbb69d7049568a671e189c4d31df99d14f7de70569d', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ebb9b42d163cdb490d062e6074b99691af113a221b62481074d5a150e8044129', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '2c477485e8286e4a0e2e6d3499c0dc253c6c858a2d7106f81cae62bb3c374eea', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'f2f4183c1750edb2a41639c237d8eb38ee30d96ee3593749616ba4f4b9c5c515', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e18421415e4ffa13dbe3a848dfdf8b91a1323c1f464879f5ec2e6aab512a0f6c', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd775a0281c84c1125a58c04446db7e74566793216c3126ce998f8163abb8bc89', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '9ef2f3ed84c16fec0eb076bb311683c068143380e4f67d2cb7d7ab8c906fed0b', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3ac4c9380a953c178f1b4d8a9ebd52716be1370e8865b0c2c7891f740028771a', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '281009e18ac6262f07d27fc06505b3524e5a3cb883aa16c0504e87f326290c63', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ac7b1afd4a8f6ca6f58ac6dc079c6e531e13b8a15c2c277555ce9c5de25c2d64', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '5b1338247b29769164d10e260f095122dca1f89297144212768f1fd6c4fb6fb7', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ff5b1b914ed4c437466f8660c32a9afbeb3695e9ac20e478f9c1ce409f1ed09e', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'bdc4ceb66ebf8c37ba5c65183f623ca08c64fe8018a102a4f0f5eabea3c15fc2', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e79d34b43243c4ec6976fac2d1089c8f2f3486f55094e31b12c811ad29a25c1c', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f3218cdcc9e6ffb5ab7cbc4e409a97c5cba656cef510c1e994aae98145b31869', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '49e8eee6f6a9043366b38d966aa36fed7ffa7a7c6fc5094b528cb58857b99f0f', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': 'c2bbcab38cdb8bcb221ca845f7476b10094b0cf3cf85313738dd3864b6053f52', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '612bc1ad6c697915f5f79c2f9ad3754ced490c89063b9bfdddf3dc5ed1737430', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '2314752bd091bb01a3c9c9f4d8f863e576f800a689d5703521d8306adffd76e2', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0a9ebc944b2de755c4512d2b1442597d5fe4b5297d62fadd6fd92902e760637f', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd155fbae7e1e3e7432c43233b801b4ce10005b36cfb24888760f62814956a602', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'abe890022cb2e1a31afca9ad0fc40511ef374822a7fcd357597aec36b3b503d0', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '5a0b086f92fe22dde447751035ee4785535bb4d9089ba71717434afd60163dc7', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'ea7350e3ce894a85dd937116b86ff58eefc778996eb3fc48d232d1c1dac27def', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:2.2.0.20251130', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 9 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'da395bf52b74e3fda92ab38bc0ddabb7e29028090e8635bdc44d7151a273550e', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20251130', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7d2f8364b0ceaf10c88bb844de62d071fc4f0a0c28a054f6da0b22c9d76e89c3', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20251130', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 10 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7a175bd7fe74190008d3ab3b9bbb0e7d06a5c2d88606abf1f5094899ee6eeadb', 'image': 'registry.osism.tech/kolla/release/ceilometer-compute:23.0.2.20251130', 'name': '/ceilometer_compute', 'state': 'running', 'status': 'Up 20 minutes (unhealthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9c0845a7a0f6ebf9eaf9aa84f6ffbeab245c4ade1433d91304b63e3259f88214', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.2.1.20251130', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 41 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b7e86c8b6f142da7eb27605a4d20d1731392d015424cc821aa9ce2b9dfb09a49', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:10.0.0.20251130', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 42 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f7d648df1e97280d44e097087185c63d2d79f38b50b281d365f017a81c42fc34', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.2.1.20251130', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 43 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8046f86788dee700330fed8eddfd693504584dd39873e897efa033060d800080', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.2.20251130', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 49 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b2877433cc1e735f486090df103fc266a4811a2c8e87200a6ec7cb1adbcafcc4', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'a31876dbc950b24a0d90116efabb4b484220e273b36b95211ea490ba31cbf16f', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c7485c82d2f1d547f6bd4f7b5f77839137922770f8e90166964bc693f5838c04', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '53b57430f3c2ea3e26a0313bd0f570ea9f96ee2cccf3420967739a15176d48d7', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '89e6fe0b290936f07e5cd92c26a65d7e4bad4aab8803b8ba0f073aa6f33518d8', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up About an hour'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'da50a7d28d968683ce4d4ca44f0cc1ebfe8cb21bf5813a11fffb890dc0b60b07', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.3.20251130', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up About an hour'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '837bfa57508f3be2114e78e4ad24b6ea1a5bad3c5b3d7e9b2025c2367a5971e5', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.3.20251130', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4d99cd8c4e4111585315968fcf3c1d2f5e140741723d97a5b22057ceeb981db5', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.3.20251130', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up About an hour (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '30b368a03fb8c09907c752cec0533ecbd1ff372b353299d4030ee6f65a37b256', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251130', 'name': '/cron', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'a6edd44b74454d1b03eb0b9925de800df0e9da371df15c735dedb32ddf11cc06', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.7.1.20251130', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'af8e0c0cd5013d2851f9115b1362788481251f73a98ac8f2abdd4398c7c1a67b', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251130', 'name': '/fluentd', 'state': 'running', 'status': 'Up 2 hours'}) \u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers on host] ********************************",
                                "Thursday 26 March 2026  04:06:17 +0000 (0:00:00.515)       0:00:05.298 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed when count of containers is wrong] *************",
                                "Thursday 26 March 2026  04:06:17 +0000 (0:00:00.296)       0:00:05.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if count matches] ******************************",
                                "Thursday 26 March 2026  04:06:18 +0000 (0:00:00.481)       0:00:06.075 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:06:18 +0000 (0:00:00.313)       0:00:06.389 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers that are not running] ********************",
                                "Thursday 26 March 2026  04:06:18 +0000 (0:00:00.304)       0:00:06.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers that are not running] *******************",
                                "Thursday 26 March 2026  04:06:19 +0000 (0:00:00.326)       0:00:07.020 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Thursday 26 March 2026  04:06:19 +0000 (0:00:00.504)       0:00:07.524 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Thursday 26 March 2026  04:06:19 +0000 (0:00:00.304)       0:00:07.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if all containers are running] *****************",
                                "Thursday 26 March 2026  04:06:20 +0000 (0:00:00.307)       0:00:08.136 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:06:20 +0000 (0:00:00.336)       0:00:08.472 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:06:21 +0000 (0:00:00.680)       0:00:09.153 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:06:21 +0000 (0:00:00.275)       0:00:09.428 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:21 +0000 (0:00:00.255)       0:00:09.683 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:21 +0000 (0:00:00.068)       0:00:09.752 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:21 +0000 (0:00:00.069)       0:00:09.821 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:06:22 +0000 (0:00:00.079)       0:00:09.901 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail early due to containers not running] ********************************",
                                "Thursday 26 March 2026  04:06:22 +0000 (0:00:00.244)       0:00:10.145 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:06:22 +0000 (0:00:00.248)       0:00:10.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set _mon_hostname fact] **************************************************",
                                "Thursday 26 March 2026  04:06:22 +0000 (0:00:00.286)       0:00:10.680 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get ceph osd tree] *******************************************************",
                                "Thursday 26 March 2026  04:06:23 +0000 (0:00:00.679)       0:00:11.360 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Parse osd tree from JSON] ************************************************",
                                "Thursday 26 March 2026  04:06:25 +0000 (0:00:01.722)       0:00:13.082 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get OSDs that are not up or in] ******************************************",
                                "Thursday 26 March 2026  04:06:25 +0000 (0:00:00.136)       0:00:13.219 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail test if OSDs are not up or in] **************************************",
                                "Thursday 26 March 2026  04:06:25 +0000 (0:00:00.318)       0:00:13.538 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Pass test if OSDs are all up and in] *************************************",
                                "Thursday 26 March 2026  04:06:25 +0000 (0:00:00.130)       0:00:13.668 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:06:25 +0000 (0:00:00.131)       0:00:13.800 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [List ceph LVM volumes and collect data] **********************************",
                                "Thursday 26 March 2026  04:06:26 +0000 (0:00:00.303)       0:00:14.103 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Parse LVM data as JSON] **************************************************",
                                "Thursday 26 March 2026  04:06:28 +0000 (0:00:02.574)       0:00:16.678 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get unencrypted and encrypted OSDs] **************************************",
                                "Thursday 26 March 2026  04:06:29 +0000 (0:00:00.326)       0:00:17.004 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of encrypted OSDs does not match] **************************",
                                "Thursday 26 March 2026  04:06:29 +0000 (0:00:00.581)       0:00:17.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of encrypted OSDs equals count of OSDs] ********************",
                                "Thursday 26 March 2026  04:06:30 +0000 (0:00:00.306)       0:00:17.892 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of unencrypted OSDs does not match] ************************",
                                "Thursday 26 March 2026  04:06:30 +0000 (0:00:00.565)       0:00:18.458 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************",
                                "Thursday 26 March 2026  04:06:30 +0000 (0:00:00.311)       0:00:18.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Thursday 26 March 2026  04:06:31 +0000 (0:00:00.297)       0:00:19.067 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get CRUSH node data of each OSD host and root node childs] ***************",
                                "Thursday 26 March 2026  04:06:31 +0000 (0:00:00.503)       0:00:19.571 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate sub test expression results] ***********************************",
                                "Thursday 26 March 2026  04:06:32 +0000 (0:00:00.776)       0:00:20.347 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail test if any sub test failed] ****************************************",
                                "Thursday 26 March 2026  04:06:32 +0000 (0:00:00.330)       0:00:20.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass test if no sub test failed] *****************************************",
                                "Thursday 26 March 2026  04:06:33 +0000 (0:00:00.339)       0:00:21.017 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Thursday 26 March 2026  04:06:33 +0000 (0:00:00.507)       0:00:21.525 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Thursday 26 March 2026  04:06:33 +0000 (0:00:00.283)       0:00:21.808 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Thursday 26 March 2026  04:06:34 +0000 (0:00:00.264)       0:00:22.073 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Thursday 26 March 2026  04:06:35 +0000 (0:00:01.680)       0:00:23.753 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Thursday 26 March 2026  04:06:36 +0000 (0:00:00.276)       0:00:24.030 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:36 +0000 (0:00:00.248)       0:00:24.279 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:36 +0000 (0:00:00.070)       0:00:24.349 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Thursday 26 March 2026  04:06:36 +0000 (0:00:00.070)       0:00:24.419 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Thursday 26 March 2026  04:06:36 +0000 (0:00:00.074)       0:00:24.494 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Thursday 26 March 2026  04:06:38 +0000 (0:00:01.557)       0:00:26.051 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-26T04:06:13+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:06:38 +0000 (0:00:00.590)       0:00:26.642 ******** ",
                                "=============================================================================== ",
                                "List ceph LVM volumes and collect data ---------------------------------- 2.57s",
                                "Get ceph osd tree ------------------------------------------------------- 1.72s",
                                "Aggregate test results step one ----------------------------------------- 1.68s",
                                "Write report file ------------------------------------------------------- 1.56s",
                                "Get timestamp for report file ------------------------------------------- 0.87s",
                                "Create report output directory ------------------------------------------ 0.79s",
                                "Calculate total number of OSDs in cluster ------------------------------- 0.78s",
                                "Get CRUSH node data of each OSD host and root node childs --------------- 0.78s",
                                "Aggregate test results step one ----------------------------------------- 0.68s",
                                "Set _mon_hostname fact -------------------------------------------------- 0.68s",
                                "Print report file information ------------------------------------------- 0.59s",
                                "Get unencrypted and encrypted OSDs -------------------------------------- 0.58s",
                                "Pass if count of encrypted OSDs equals count of OSDs -------------------- 0.57s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.55s",
                                "Get list of ceph-osd containers on host --------------------------------- 0.52s",
                                "Pass test if no sub test failed ----------------------------------------- 0.51s",
                                "Get count of ceph-osd containers that are not running ------------------- 0.50s",
                                "Prepare test data ------------------------------------------------------- 0.50s",
                                "Set test result to failed when count of containers is wrong ------------- 0.48s",
                                "Fail test if any sub test failed ---------------------------------------- 0.34s",
                                "+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ source /etc/os-release",
                                "++ PRETTY_NAME='Ubuntu 24.04.4 LTS'",
                                "++ NAME=Ubuntu",
                                "++ VERSION_ID=24.04",
                                "++ VERSION='24.04.4 LTS (Noble Numbat)'",
                                "++ VERSION_CODENAME=noble",
                                "++ ID=ubuntu",
                                "++ ID_LIKE=debian",
                                "++ HOME_URL=https://www.ubuntu.com/",
                                "++ SUPPORT_URL=https://help.ubuntu.com/",
                                "++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/",
                                "++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy",
                                "++ UBUNTU_CODENAME=noble",
                                "++ LOGO=ubuntu-logo",
                                "+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]",
                                "+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'",
                                "+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "",
                                "# Status of Elasticsearch",
                                "",
                                "+ pushd /opt/configuration/contrib",
                                "+ echo",
                                "+ echo '# Status of Elasticsearch'",
                                "+ echo",
                                "+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s",
                                "OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0",
                                "",
                                "# Status of MariaDB",
                                "",
                                "+ echo",
                                "+ echo '# Status of MariaDB'",
                                "+ echo",
                                "++ semver 9.5.0 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]",
                                "+ MARIADB_USER=root_shard_0",
                                "+ bash nagios-plugins/check_galera_cluster -u root_shard_0 -p password -H api-int.testbed.osism.xyz -c 1",
                                "Reading package lists...",
                                "Building dependency tree...",
                                "Reading state information...",
                                "bc is already the newest version (1.07.1-3ubuntu4).",
                                "bc set to manually installed.",
                                "0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.",
                                "OK: number of NODES = 3 (wsrep_cluster_size)",
                                "",
                                "# Status of Prometheus",
                                "",
                                "+ echo",
                                "+ echo '# Status of Prometheus'",
                                "+ echo",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy",
                                "Unauthorized",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready",
                                "Unauthorized",
                                "",
                                "# Status of RabbitMQ",
                                "",
                                "+ echo",
                                "+ echo '# Status of RabbitMQ'",
                                "+ echo",
                                "++ semver 9.5.0 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]",
                                "+ perl nagios-plugins/check_rabbitmq_cluster --ssl 1 -H api-int.testbed.osism.xyz -u openstack -p password",
                                "RABBITMQ_CLUSTER OK - nb_running_node OK (3) nb_running_disc_node OK (3) nb_running_ram_node OK (0)",
                                "",
                                "# Status of Redis",
                                "",
                                "+ echo",
                                "+ echo '# Status of Redis'",
                                "+ echo",
                                "+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j",
                                "TCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.002068s;;;0.000000;10.000000",
                                "",
                                "+ popd",
                                "+ echo",
                                "# Create backup of MariaDB database",
                                "",
                                "+ echo '# Create backup of MariaDB database'",
                                "+ echo",
                                "+ osism apply mariadb_backup -e mariadb_backup_type=full",
                                "\u001b[32m2026-03-26 04:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7b1aa3d5-846c-4962-a0bc-63f2776dc6a9 (mariadb_backup) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7b1aa3d5-846c-4962-a0bc-63f2776dc6a9 (mariadb_backup) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:07:11 +0000 (0:00:00.177)       0:00:00.177 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:07:11 +0000 (0:00:00.313)       0:00:00.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Thursday 26 March 2026  04:07:12 +0000 (0:00:00.652)       0:00:01.144 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  04:07:12 +0000 (0:00:00.429)       0:00:01.574 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB container facts] ***********************************",
                                "Thursday 26 March 2026  04:07:13 +0000 (0:00:00.572)       0:00:02.147 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Taking full database backup via Mariabackup] *******************",
                                "Thursday 26 March 2026  04:07:16 +0000 (0:00:03.248)       0:00:05.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "STILL ALIVE [task 'mariadb : Taking full database backup via Mariabackup' is running] ***",
                                "",
                                "STILL ALIVE [task 'mariadb : Taking full database backup via Mariabackup' is running] ***",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Thursday 26 March 2026  04:10:25 +0000 (0:03:08.742)       0:03:14.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Thursday 26 March 2026  04:10:25 +0000 (0:00:00.304)       0:03:14.442 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:10:26 +0000 (0:00:00.435)       0:03:14.878 ******** ",
                                "=============================================================================== ",
                                "mariadb : Taking full database backup via Mariabackup ----------------- 188.74s",
                                "mariadb : Get MariaDB container facts ----------------------------------- 3.25s",
                                "Group hosts based on enabled services ----------------------------------- 0.65s",
                                "mariadb : include_tasks ------------------------------------------------- 0.57s",
                                "Include mariadb post-upgrade.yml ---------------------------------------- 0.44s",
                                "mariadb : Group MariaDB hosts based on shards --------------------------- 0.43s",
                                "Group hosts based on Kolla action --------------------------------------- 0.31s",
                                "Include mariadb post-deploy.yml ----------------------------------------- 0.30s",
                                "+ sh -c /opt/configuration/scripts/check/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# OpenStack endpoints",
                                "",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "+ export OS_CLOUD=admin",
                                "+ OS_CLOUD=admin",
                                "+ echo",
                                "+ echo '# OpenStack endpoints'",
                                "+ echo",
                                "+ openstack endpoint list",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| 0033fc892da9412fa8a093fd5d2a963f | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |",
                                "| 08783899140940f1ab1e28d0eb212655 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |",
                                "| 0a595fefe7e94bd7a92b4bec73d647a6 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |",
                                "| 186cae91475646dfbbc932451c78d918 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |",
                                "| 2070bf277fa04949b94df4bb73a01ad5 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |",
                                "| 214d96bfc5c74deaa0287323a323b818 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |",
                                "| 3089431d8d8b4f0ba7824a5faf669f60 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |",
                                "| 35fc90d9137b4d7f92d7ac68014b9fc6 | RegionOne | manila       | share           | True    | internal  | https://api-int.testbed.osism.xyz:8786/v1/%(tenant_id)s             |",
                                "| 3b15fd0003e04f9f8aa0c22a2ea26c73 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |",
                                "| 4fbec3dc7ce74b8cb3c6d0974f2e9e1d | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |",
                                "| 52db55a577c242bea758b13e5034c75a | RegionOne | aodh         | alarming        | True    | public    | https://api.testbed.osism.xyz:8042                                  |",
                                "| 56d8259ac37946e2aff58d5495dc0f81 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |",
                                "| 57fa80cce7544a10a586776995a291a6 | RegionOne | skyline      | panel           | True    | internal  | https://api-int.testbed.osism.xyz:9998                              |",
                                "| 75c163e497a14d5bb1f2b9ee146083e9 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |",
                                "| 856b24feafb3443b9aaeae1f19b86647 | RegionOne | manilav2     | sharev2         | True    | internal  | https://api-int.testbed.osism.xyz:8786/v2                           |",
                                "| 8ba59f40ae3b4193b5819b7c08fd3243 | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |",
                                "| 9809bb157e744c398e8f9c7fcac26a1a | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |",
                                "| 9989f77bad6840df9a22ad29181f93ea | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |",
                                "| a59948b5184e45e084d6aaa31149f210 | RegionOne | skyline      | panel           | True    | public    | https://api.testbed.osism.xyz:9998                                  |",
                                "| a5d9ce36abcb47c3b75e5988e8e015df | RegionOne | manilav2     | sharev2         | True    | public    | https://api.testbed.osism.xyz:8786/v2                               |",
                                "| b53ac5aaf5914c0cb65e717d7642ad90 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |",
                                "| c577ae50a66d49879af93d9946c176b8 | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |",
                                "| c8cf89a51f61454891009bb19cafd4c4 | RegionOne | aodh         | alarming        | True    | internal  | https://api-int.testbed.osism.xyz:8042                              |",
                                "| ce290ff5801e446f8041086f7a131bab | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |",
                                "| d4f79ca24681457c8ddb3d19e9973a63 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |",
                                "| db3adab6cd89404792a59d0f59461602 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |",
                                "| db945c2decd04c0bb50c44206ee77a79 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |",
                                "| e3fa99a3313a44d1acb8aa7427f045de | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |",
                                "| f1d018f3e5f6498c9420138b7ade2a36 | RegionOne | manila       | share           | True    | public    | https://api.testbed.osism.xyz:8786/v1/%(tenant_id)s                 |",
                                "| f7da6a40e78f462ea2d8069b207f9e0d | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "",
                                "# Cinder",
                                "",
                                "+ echo",
                                "+ echo '# Cinder'",
                                "+ echo",
                                "+ openstack volume service list",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| Binary           | Host                       | Zone     | Status  | State | Updated At                 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |",
                                "| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |",
                                "| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-26T04:10:24.000000 |",
                                "| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:24.000000 |",
                                "| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:31.000000 |",
                                "| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-26T04:10:32.000000 |",
                                "| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-26T04:10:29.000000 |",
                                "| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-26T04:10:31.000000 |",
                                "| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-26T04:10:32.000000 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "",
                                "# Neutron",
                                "",
                                "+ echo",
                                "+ echo '# Neutron'",
                                "+ echo",
                                "+ openstack network agent list",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |",
                                "| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "+ openstack network service provider list",
                                "+---------------+------+---------+",
                                "| Service Type  | Name | Default |",
                                "+---------------+------+---------+",
                                "| L3_ROUTER_NAT | ovn  | True    |",
                                "+---------------+------+---------+",
                                "",
                                "# Nova",
                                "",
                                "+ echo",
                                "+ echo '# Nova'",
                                "+ echo",
                                "+ openstack compute service list",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| 86eb2092-fe79-4181-8a5e-f8a408c1e3b9 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-26T04:10:40.000000 |",
                                "| dd2eeb5e-c5d1-428b-bbce-901a429360a4 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-26T04:10:35.000000 |",
                                "| 5892028d-cfad-462c-9137-727335627c04 | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-26T04:10:37.000000 |",
                                "| f48bec24-2526-47fb-89c7-f4540915a0fa | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-26T04:10:37.000000 |",
                                "| 7a41d58a-be90-4c07-b6b6-e07900ee5ad4 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-26T04:10:39.000000 |",
                                "| f0aa1fcd-e220-419e-887b-6a41fe3dd532 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-26T04:10:40.000000 |",
                                "| 83b43651-bd92-4d18-a2b5-fcbdc87357c5 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |",
                                "| b9e61882-f01c-4632-a3ea-b296d9feba1c | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |",
                                "| eff6a403-1dbd-4531-b082-376def61e6f9 | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-26T04:10:35.000000 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "+ openstack hypervisor list",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| 227f7a23-f93c-4b92-b429-5a83b73d9fd5 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |",
                                "| 147427a8-4597-4a2c-823c-3005e03af22a | testbed-node-3      | QEMU            | 192.168.16.13 | up    |",
                                "| 4a11d35e-b357-4267-8292-b9149dc34db0 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "+ echo",
                                "",
                                "# Run OpenStack test play",
                                "",
                                "+ echo '# Run OpenStack test play'",
                                "+ echo",
                                "+ osism apply --environment openstack test",
                                "\u001b[32m2026-03-26 04:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-26 04:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 49017abc-3eae-4005-afad-1c43fa7ba55f (test) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 49017abc-3eae-4005-afad-1c43fa7ba55f (test) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Create test project] *****************************************************",
                                "",
                                "TASK [Create test domain] ******************************************************",
                                "Thursday 26 March 2026  04:11:01 +0000 (0:00:00.080)       0:00:00.080 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test-admin user] **************************************************",
                                "Thursday 26 March 2026  04:11:04 +0000 (0:00:03.755)       0:00:03.835 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add manager role to user test-admin] *************************************",
                                "Thursday 26 March 2026  04:11:09 +0000 (0:00:04.139)       0:00:07.975 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test project] *****************************************************",
                                "Thursday 26 March 2026  04:11:15 +0000 (0:00:06.524)       0:00:14.499 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test user] ********************************************************",
                                "Thursday 26 March 2026  04:11:19 +0000 (0:00:03.974)       0:00:18.474 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add member roles to user test] *******************************************",
                                "Thursday 26 March 2026  04:11:23 +0000 (0:00:04.185)       0:00:22.659 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m",
                                "",
                                "TASK [Create test server group] ************************************************",
                                "Thursday 26 March 2026  04:11:35 +0000 (0:00:11.421)       0:00:34.080 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create ssh security group] ***********************************************",
                                "Thursday 26 March 2026  04:11:40 +0000 (0:00:05.096)       0:00:39.177 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to ssh security group] ******************************************",
                                "Thursday 26 March 2026  04:11:45 +0000 (0:00:04.925)       0:00:44.102 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create icmp security group] **********************************************",
                                "Thursday 26 March 2026  04:11:49 +0000 (0:00:04.320)       0:00:48.423 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to icmp security group] *****************************************",
                                "Thursday 26 March 2026  04:11:53 +0000 (0:00:03.923)       0:00:52.347 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test keypair] *****************************************************",
                                "Thursday 26 March 2026  04:11:57 +0000 (0:00:04.091)       0:00:56.438 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test network] *****************************************************",
                                "Thursday 26 March 2026  04:12:01 +0000 (0:00:03.894)       0:01:00.332 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test subnet] ******************************************************",
                                "Thursday 26 March 2026  04:12:06 +0000 (0:00:04.908)       0:01:05.241 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test router] ******************************************************",
                                "Thursday 26 March 2026  04:12:12 +0000 (0:00:05.809)       0:01:11.050 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Manage test instances and volumes] ***************************************",
                                "",
                                "TASK [Get test server group] ***************************************************",
                                "Thursday 26 March 2026  04:12:23 +0000 (0:00:11.547)       0:01:22.598 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Detach test volume] ******************************************************",
                                "Thursday 26 March 2026  04:12:27 +0000 (0:00:03.550)       0:01:26.149 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test volume] ******************************************************",
                                "Thursday 26 March 2026  04:12:27 +0000 (0:00:00.058)       0:01:26.207 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test instances] ***************************************************",
                                "Thursday 26 March 2026  04:12:27 +0000 (0:00:00.055)       0:01:26.263 ******** ",
                                "\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Wait for instance deletion to complete] **********************************",
                                "Thursday 26 March 2026  04:12:27 +0000 (0:00:00.157)       0:01:26.420 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test instances] ***************************************************",
                                "Thursday 26 March 2026  04:12:27 +0000 (0:00:00.152)       0:01:26.572 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for instance creation to complete] **********************************",
                                "Thursday 26 March 2026  04:12:32 +0000 (0:00:04.784)       0:01:31.357 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j538260717191.3753', 'results_file': '/ansible/.ansible_async/j538260717191.3753', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j899561745004.3778', 'results_file': '/ansible/.ansible_async/j899561745004.3778', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j588604499832.3803', 'results_file': '/ansible/.ansible_async/j588604499832.3803', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j405640357128.3828', 'results_file': '/ansible/.ansible_async/j405640357128.3828', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j525201259332.3853', 'results_file': '/ansible/.ansible_async/j525201259332.3853', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add metadata to instances] ***********************************************",
                                "Thursday 26 March 2026  04:13:29 +0000 (0:00:57.245)       0:02:28.602 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for metadata to be added] *******************************************",
                                "Thursday 26 March 2026  04:13:34 +0000 (0:00:04.390)       0:02:32.993 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j984712504400.3965', 'results_file': '/ansible/.ansible_async/j984712504400.3965', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j533774217712.3990', 'results_file': '/ansible/.ansible_async/j533774217712.3990', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j898183755290.4015', 'results_file': '/ansible/.ansible_async/j898183755290.4015', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j768463755450.4040', 'results_file': '/ansible/.ansible_async/j768463755450.4040', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j698557361688.4065', 'results_file': '/ansible/.ansible_async/j698557361688.4065', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add tag to instances] ****************************************************",
                                "Thursday 26 March 2026  04:13:43 +0000 (0:00:09.467)       0:02:42.461 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for tags to be added] ***********************************************",
                                "Thursday 26 March 2026  04:13:48 +0000 (0:00:04.613)       0:02:47.075 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j766959703281.4134', 'results_file': '/ansible/.ansible_async/j766959703281.4134', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j744976519201.4159', 'results_file': '/ansible/.ansible_async/j744976519201.4159', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j468878930547.4191', 'results_file': '/ansible/.ansible_async/j468878930547.4191', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j481909587613.4217', 'results_file': '/ansible/.ansible_async/j481909587613.4217', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j4746500476.4243', 'results_file': '/ansible/.ansible_async/j4746500476.4243', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Create test volume] ******************************************************",
                                "Thursday 26 March 2026  04:13:57 +0000 (0:00:09.731)       0:02:56.806 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Attach test volume] ******************************************************",
                                "Thursday 26 March 2026  04:14:04 +0000 (0:00:06.098)       0:03:02.904 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create floating ip address] **********************************************",
                                "Thursday 26 March 2026  04:14:17 +0000 (0:00:13.445)       0:03:16.350 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Print floating ip address] ***********************************************",
                                "Thursday 26 March 2026  04:14:22 +0000 (0:00:05.342)       0:03:21.693 ******** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"192.168.112.156\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:14:22 +0000 (0:00:00.050)       0:03:21.744 ******** ",
                                "=============================================================================== ",
                                "Wait for instance creation to complete --------------------------------- 57.25s",
                                "Attach test volume ----------------------------------------------------- 13.45s",
                                "Create test router ----------------------------------------------------- 11.55s",
                                "Add member roles to user test ------------------------------------------ 11.42s",
                                "Wait for tags to be added ----------------------------------------------- 9.73s",
                                "Wait for metadata to be added ------------------------------------------- 9.47s",
                                "Add manager role to user test-admin ------------------------------------- 6.52s",
                                "Create test volume ------------------------------------------------------ 6.10s",
                                "Create test subnet ------------------------------------------------------ 5.81s",
                                "Create floating ip address ---------------------------------------------- 5.34s",
                                "Create test server group ------------------------------------------------ 5.10s",
                                "Create ssh security group ----------------------------------------------- 4.93s",
                                "Create test network ----------------------------------------------------- 4.91s",
                                "Create test instances --------------------------------------------------- 4.78s",
                                "Add tag to instances ---------------------------------------------------- 4.61s",
                                "Add metadata to instances ----------------------------------------------- 4.39s",
                                "Add rule to ssh security group ------------------------------------------ 4.32s",
                                "Create test user -------------------------------------------------------- 4.19s",
                                "Create test-admin user -------------------------------------------------- 4.14s",
                                "Add rule to icmp security group ----------------------------------------- 4.09s",
                                "+ server_list",
                                "+ openstack --os-cloud test server list",
                                "+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+",
                                "| ID                                   | Name   | Status | Networks                             | Image                    | Flavor   |",
                                "+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+",
                                "| 9f84049c-743f-4895-8425-3c6d6b59650a | test-4 | ACTIVE | test=192.168.112.180, 192.168.200.63 | N/A (booted from volume) | SCS-1L-1 |",
                                "| a86d3bc5-e53a-421d-bf49-90bac35835eb | test-3 | ACTIVE | test=192.168.112.146, 192.168.200.91 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 72fea215-b4a0-47d7-a5f6-1c164d423e5f | test   | ACTIVE | test=192.168.112.156, 192.168.200.3  | N/A (booted from volume) | SCS-1L-1 |",
                                "| f05d3527-1651-466e-b2c0-b2e9034c1f08 | test-2 | ACTIVE | test=192.168.112.181, 192.168.200.17 | N/A (booted from volume) | SCS-1L-1 |",
                                "| f6a0beab-2de6-40d7-8e1c-30e7df59b6ae | test-1 | ACTIVE | test=192.168.112.117, 192.168.200.89 | N/A (booted from volume) | SCS-1L-1 |",
                                "+--------------------------------------+--------+--------+--------------------------------------+--------------------------+----------+",
                                "+ openstack --os-cloud test server show test",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-26T04:13:03.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.156, 192.168.200.3                                                                                                                                                                                                                                                                                                                                                 |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 3c4a154524be0044d9af6f6e2ac04bbe090637876489341638c46842                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 72fea215-b4a0-47d7-a5f6-1c164d423e5f                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='41d770c6-35d8-40d7-98a6-bf5de323a1d4'                                                                                                                                                                                                                                                                                                             |",
                                "|                                     | delete_on_termination='False', id='c0b1f93f-9c66-4f28-98bf-725944fcdd59'                                                                                                                                                                                                                                                                                                            |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-1",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-26T04:13:05.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.117, 192.168.200.89                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 3c4a154524be0044d9af6f6e2ac04bbe090637876489341638c46842                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | f6a0beab-2de6-40d7-8e1c-30e7df59b6ae                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='8241a54e-f4c8-4656-ba7f-0cc497b28890'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-2",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-26T04:13:04.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.181, 192.168.200.17                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-26T04:12:37Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 674ca42258a0b4c8c8031bac2c2e2fbf6c22c7834e3bc95678cd74b8                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | f05d3527-1651-466e-b2c0-b2e9034c1f08                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-26T04:13:36Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='28433d04-ce59-4919-8a32-660a14b3fb94'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-3",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-26T04:13:06.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.146, 192.168.200.91                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-26T04:12:38Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 674ca42258a0b4c8c8031bac2c2e2fbf6c22c7834e3bc95678cd74b8                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | a86d3bc5-e53a-421d-bf49-90bac35835eb                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-26T04:13:37Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='7ff6c181-375d-43be-8b24-6e9effb98e3b'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-4",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-26T04:13:06.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.180, 192.168.200.63                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-26T04:12:40Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | ee0f370abe6f4901d5b81d816f3256dda7c87709752fbf58b38a05ee                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 9f84049c-743f-4895-8425-3c6d6b59650a                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 77d2c6f1cc6541f3bbb616e01134d7dd                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-26T04:13:38Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | f5786063540c4b40a3a24028a0f17712                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='2af11777-ad16-40e3-aa06-7eb5dd4b823c'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.146",
                                "PING 192.168.112.146 (192.168.112.146) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.146: icmp_seq=1 ttl=63 time=7.13 ms",
                                "64 bytes from 192.168.112.146: icmp_seq=2 ttl=63 time=2.30 ms",
                                "64 bytes from 192.168.112.146: icmp_seq=3 ttl=63 time=1.91 ms",
                                "",
                                "--- 192.168.112.146 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.905/3.779/7.133/2.376 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=8.63 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.41 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.84 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.836/4.292/8.629/3.075 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.181",
                                "PING 192.168.112.181 (192.168.112.181) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.181: icmp_seq=1 ttl=63 time=5.35 ms",
                                "64 bytes from 192.168.112.181: icmp_seq=2 ttl=63 time=2.32 ms",
                                "64 bytes from 192.168.112.181: icmp_seq=3 ttl=63 time=2.04 ms",
                                "",
                                "--- 192.168.112.181 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.039/3.237/5.352/1.499 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.77 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.27 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.64 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.641/4.227/8.773/3.224 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.180",
                                "PING 192.168.112.180 (192.168.112.180) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.180: icmp_seq=1 ttl=63 time=7.43 ms",
                                "64 bytes from 192.168.112.180: icmp_seq=2 ttl=63 time=2.92 ms",
                                "64 bytes from 192.168.112.180: icmp_seq=3 ttl=63 time=1.78 ms",
                                "",
                                "--- 192.168.112.180 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.782/4.043/7.431/2.439 ms",
                                "+ [[ 9.5.0 == \\l\\a\\t\\e\\s\\t ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000001f-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:57.972364Z",
                            "start": "2026-03-26T04:04:05.449657Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000001f",
                        "name": "Run checks"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_tempest | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:58.552920Z",
                            "start": "2026-03-26T04:14:58.042113Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000020",
                        "name": "Run tempest"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_prometheus_alert_status | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.107669Z",
                            "start": "2026-03-26T04:14:58.596716Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000021",
                        "name": "Check prometheus alert status"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:19:28.816979Z",
                    "start": "2026-03-26T04:14:59.121212Z"
                },
                "id": "0242ac1f-6606-ba6a-269d-000000000022",
                "name": "Upgrade testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version_next": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.214286Z",
                            "start": "2026-03-26T04:14:59.132678Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000024",
                        "name": "Print next ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version_next": "2024.2",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.306299Z",
                            "start": "2026-03-26T04:14:59.224941Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000025",
                        "name": "Print next openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version_next": "10.0.0-rc.1",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.387156Z",
                            "start": "2026-03-26T04:14:59.318104Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000026",
                        "name": "Print next manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "kolla_namespace": "kolla/release"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.456535Z",
                            "start": "2026-03-26T04:14:59.403215Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000027",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.505279Z",
                            "start": "2026-03-26T04:14:59.467740Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000028",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTIuNTQK",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.806664Z",
                            "start": "2026-03-26T04:14:59.521490Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-000000000029",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.192.54"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T04:14:59.897519Z",
                            "start": "2026-03-26T04:14:59.816868Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000002a",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.54",
                                "/opt/configuration/scripts/upgrade.sh",
                                "10.0.0-rc.1",
                                "reef",
                                "2024.2",
                                "kolla/release"
                            ],
                            "delta": "2:04:28.073244",
                            "end": "2026-03-26 06:19:28.265674",
                            "failed": true,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.54 /opt/configuration/scripts/upgrade.sh 10.0.0-rc.1 reef 2024.2 kolla/release",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000002b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 2,
                            "start": "2026-03-26 04:15:00.192430",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ export MANAGER_VERSION=10.0.0-rc.1\n+ MANAGER_VERSION=10.0.0-rc.1\n+ CEPH_VERSION=reef\n+ OPENSTACK_VERSION=2024.2\n+ KOLLA_NAMESPACE=kolla/release\n+ sh -c '/opt/configuration/scripts/upgrade-manager.sh 10.0.0-rc.1 reef 2024.2 kolla/release'\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n++ docker inspect --format '{{ index .Config.Labels \"org.opencontainers.image.version\"}}' osism-ansible\n+ OLD_MANAGER_VERSION=v0.20251130.0\n++ docker inspect --format '{{ index .Config.Labels \"de.osism.release.openstack\"}}' kolla-ansible\n\n# UPGRADE MANAGER\n\n+ OLD_OPENSTACK_VERSION=2024.2\n+ echo\n+ echo '# UPGRADE MANAGER'\n+ echo\n+ export MANAGER_VERSION=10.0.0-rc.1\n+ MANAGER_VERSION=10.0.0-rc.1\n+ CEPH_VERSION=reef\n+ OPENSTACK_VERSION=2024.2\n+ KOLLA_NAMESPACE=kolla/release\n+ /opt/configuration/scripts/set-manager-version.sh 10.0.0-rc.1\n+ set -e\n+ VERSION=10.0.0-rc.1\n+ sed -i 's/manager_version: .*/manager_version: 10.0.0-rc.1/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ 10.0.0-rc.1 != \\l\\a\\t\\e\\s\\t ]]\n+ sed -i /ceph_version:/d /opt/configuration/environments/manager/configuration.yml\n+ sed -i /openstack_version:/d /opt/configuration/environments/manager/configuration.yml\n+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh\n/opt/configuration ~\n+ set -e\n+ pushd /opt/configuration\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging\nRequirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)\nRequirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)\nRequirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)\nRequirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)\nRequirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)\nRequirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)\nRequirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)\nRequirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)\nRequirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)\nRequirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)\nRequirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)\nRequirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)\nRequirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)\nRequirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)\n++ which gilt\n+ GILT=/opt/venv/bin/gilt\n+ /opt/venv/bin/gilt overlay\nosism.cfg-generics:\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/\n  - running `rm render-images.py` in /opt/configuration/environments/manager/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/\n  - running `rm set-versions.py` in /opt/configuration/environments/\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ popd\n~\n+ [[ 10.0.0-rc.1 == \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release\n+ set -e\n+ NAMESPACE=kolla/release\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh\n/opt/configuration ~\n+ set -e\n+ pushd /opt/configuration\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging\nRequirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)\nRequirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)\nRequirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)\nRequirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)\nRequirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)\nRequirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)\nRequirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)\nRequirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)\nRequirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)\nRequirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)\nRequirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)\nRequirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)\nRequirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)\nRequirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)\n++ which gilt\n+ GILT=/opt/venv/bin/gilt\n+ /opt/venv/bin/gilt overlay\nosism.cfg-generics:\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/\n  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/\n  - running `rm render-images.py` in /opt/configuration/environments/manager/\n  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/\n  - running `rm set-versions.py` in /opt/configuration/environments/\n~\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ popd\n++ semver v0.20251130.0 6.0.0\n+ [[ -1 -ge 0 ]]\n++ semver 10.0.0-rc.1 10.0.0-0\n+ [[ 1 -ge 0 ]]\n+ sed -i '/^om_enable_rabbitmq_high_availability:/d' /opt/configuration/environments/kolla/configuration.yml\n+ sed -i '/^om_enable_rabbitmq_quorum_queues:/d' /opt/configuration/environments/kolla/configuration.yml\n+++ semver v0.20251130.0 9.5.0\n++ '[' -1 -le 0 ']'\n+++ semver 10.0.0-rc.1 10.0.0-0\n++ '[' 1 -ge 0 ']'\n++ echo true\n+ MANAGER_UPGRADE_CROSSES_10=true\n+++ semver 2024.2 2024.2\n++ '[' 0 -le 0 ']'\n+++ semver 2024.2 2025.1\n++ '[' -1 -ge 0 ']'\n++ echo false\n+ OPENSTACK_UPGRADE_CROSSES_2025=false\n+ [[ true == \\t\\r\\u\\e ]]\n+ echo 'om_rpc_vhost: openstack'\n+ echo 'om_notify_vhost: openstack'\n+ sed -i 's#manager_listener_broker_vhost: .*#manager_listener_broker_vhost: /openstack#g' /opt/configuration/environments/manager/configuration.yml\n+ echo 'export RABBITMQ3TO4=true'\n+ sudo tee -a /opt/manager-vars.sh\nexport RABBITMQ3TO4=true\n+ osism update manager\nCollecting uv\n  Downloading uv-0.11.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB)\nDownloading uv-0.11.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (24.5 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 24.5/24.5 MB 33.8 MB/s eta 0:00:00\nInstalling collected packages: uv\nSuccessfully installed uv-0.11.1\nResolved 11 packages in 380ms\nDownloading netaddr (2.2MiB)\nDownloading cryptography (4.3MiB)\nDownloading ansible-core (2.1MiB)\nDownloading ansible (54.5MiB)\n Downloaded netaddr\n Downloaded cryptography\n Downloaded ansible-core\n Downloaded ansible\nPrepared 11 packages in 7.40s\nInstalled 11 packages in 604ms\n + ansible==11.11.0\n + ansible-core==2.18.15\n + cffi==2.0.0\n + cryptography==46.0.6\n + jinja2==3.1.6\n + markupsafe==3.0.3\n + netaddr==1.3.0\n + packaging==26.0\n + pycparser==3.0\n + pyyaml==6.0.3\n + resolvelib==1.0.1\nCloning into '/home/dragon/.ansible/tmp/ansible-local-203816n2qsv95p/tmpz7vqf5eh/ansible-collection-servicesarne1aad'...\nYour branch is up to date with 'origin/main'.\nAlready on 'main'\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/dragon/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/dragon/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully\nCloning into '/home/dragon/.ansible/tmp/ansible-local-203833xiam2p3s/tmp84azf_u6/ansible-playbooks-managerqcge2toh'...\nAlready on 'main'\nYour branch is up to date with 'origin/main'.\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.manager:999.0.0' to '/home/dragon/.ansible/collections/ansible_collections/osism/manager'\nCreated collection for osism.manager:999.0.0 at /home/dragon/.ansible/collections/ansible_collections/osism/manager\nosism.manager:999.0.0 was installed successfully\n[WARNING]: Invalid characters were found in group names but not replaced, use\n-vvvv to see details\n\nPLAY [Apply role manager] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nok: [testbed-manager] => (item=/opt/ansible)\nok: [testbed-manager] => (item=/opt/archive)\nok: [testbed-manager] => (item=/opt/manager/configuration)\nok: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nok: [testbed-manager] => (item=/opt/manager/secrets)\nok: [testbed-manager] => (item=/opt/ansible/secrets)\nok: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nok: [testbed-manager] => (item=ara)\nok: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nok: [testbed-manager] => (item=None)\nok: [testbed-manager] => (item=None)\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nok: [testbed-manager] => (item=conductor)\nok: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nok: [testbed-manager] => (item=osism)\nok: [testbed-manager] => (item=osism-update-docker)\nok: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nok: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Register that manager service was restarted] ***\nok: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=51   changed=9    unreachable=0    failed=0    skipped=8    rescued=0    ignored=0   \n\n\u001b[32m2026-03-26 04:20:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c34b26f2-f289-4f95-835a-bf861cbcec99 (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-03-26 04:20:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-03-26 04:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-03-26 04:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-03-26 04:20:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-03-26 04:20:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n+ [[ '' == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                         COMMAND                  SERVICE                CREATED         STATUS                          PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:0.20251208.0           \"/entrypoint.sh osis\u2026\"   ceph-ansible           3 minutes ago   Up 2 minutes (healthy)          \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:0.20251208.0          \"/entrypoint.sh osis\u2026\"   kolla-ansible          3 minutes ago   Up 2 minutes (healthy)          \nmanager-api-1                    registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   api                    3 minutes ago   Up 3 minutes (healthy)          192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                    \"sh -c '/wait && /ru\u2026\"   ara-server             2 hours ago     Up 2 minutes (healthy)          8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   beat                   3 minutes ago   Up 3 minutes (healthy)          \nmanager-flower-1                 registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   flower                 3 minutes ago   Up 3 minutes (healthy)          \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:0.20251208.0   \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   3 minutes ago   Up 2 minutes (healthy)          \nmanager-listener-1               registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   listener               3 minutes ago   Restarting (0) 21 seconds ago   \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4          \"docker-entrypoint.s\u2026\"   mariadb                2 hours ago     Up 3 minutes (healthy)          3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   openstack              3 minutes ago   Up 3 minutes (healthy)          \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine      \"docker-entrypoint.s\u2026\"   redis                  2 hours ago     Up 3 minutes (healthy)          6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:0.20251208.0          \"/entrypoint.sh osis\u2026\"   osism-ansible          3 minutes ago   Up 2 minutes (healthy)          \nosism-frontend                   registry.osism.tech/osism/osism-frontend:0.20251208.0         \"docker-entrypoint.s\u2026\"   frontend               3 minutes ago   Up 3 minutes                    192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:0.20251208.0       \"/entrypoint.sh osis\u2026\"   osism-kubernetes       3 minutes ago   Up 2 minutes (healthy)          \nosismclient                      registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- sleep\u2026\"   osismclient            3 minutes ago   Up 3 minutes (healthy)          \n+ [[ '' == \\t\\r\\u\\e ]]\n+ [[ '' == \\f\\a\\l\\s\\e ]]\n+ osism apply facts\n\u001b[32m2026-03-26 04:20:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe19af64-9cef-441d-8bc1-20715ae45ad3 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:20:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fe19af64-9cef-441d-8bc1-20715ae45ad3 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nThursday 26 March 2026  04:20:49 +0000 (0:00:01.971)       0:00:01.971 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nThursday 26 March 2026  04:20:52 +0000 (0:00:03.524)       0:00:05.496 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nThursday 26 March 2026  04:20:55 +0000 (0:00:02.580)       0:00:08.077 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nThursday 26 March 2026  04:21:02 +0000 (0:00:07.088)       0:00:15.165 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:21:05 +0000 (0:00:02.682)       0:00:17.848 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 7.09s\nosism.commons.facts : Create custom facts directory --------------------- 3.53s\nGather facts for all hosts ---------------------------------------------- 2.68s\nosism.commons.facts : Copy fact files ----------------------------------- 2.58s\n++ semver 10.0.0-rc.1 10.0.0-0\n+ [[ 1 -ge 0 ]]\n++ docker inspect --format '{{ index .Config.Labels \"de.osism.release.openstack\"}}' kolla-ansible\n+ OPENSTACK_VERSION=2025.1\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release/2025.1\n+ set -e\n+ NAMESPACE=kolla/release/2025.1\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release/2025.1#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n+ sh -c /opt/configuration/scripts/upgrade-services.sh\n\n# UPGRADE SERVICES\n\n+ set -e\n+ echo\n+ echo '# UPGRADE SERVICES'\n+ echo\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n++ export RABBITMQ3TO4=true\n++ RABBITMQ3TO4=true\n+ SKIP_OPENSTACK_UPGRADE=false\n+ SKIP_CEPH_UPGRADE=false\n+ sh -c /opt/configuration/scripts/pull-images.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n\n# PULL IMAGES\n\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n++ export RABBITMQ3TO4=true\n++ RABBITMQ3TO4=true\n+ echo\n+ echo '# PULL IMAGES'\n+ echo\n++ semver 9.5.0 7.0.0\n+ [[ 1 -ge 0 ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-03-26 04:21:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-03-26 04:21:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 411ad312-6a16-41aa-83a0-93c728c3fae0 (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:21:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 411ad312-6a16-41aa-83a0-93c728c3fae0 is running in background. No more output. Check ARA for logs.\u001b[0m\n+ sh -c /opt/configuration/scripts/upgrade/500-kubernetes.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n++ export MANAGER_VERSION=10.0.0-rc.1\n++ MANAGER_VERSION=10.0.0-rc.1\n++ semver 10.0.0-rc.1 8.0.3\n+ [[ 1 -ge 0 ]]\n+ osism apply frr\n\u001b[32m2026-03-26 04:21:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d394b63-d017-4309-be34-0efdf0caac8b (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:21:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0d394b63-d017-4309-be34-0efdf0caac8b (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nThursday 26 March 2026  04:21:38 +0000 (0:00:02.965)       0:00:02.965 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nThursday 26 March 2026  04:21:40 +0000 (0:00:02.108)       0:00:05.074 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nThursday 26 March 2026  04:21:42 +0000 (0:00:02.058)       0:00:07.133 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nThursday 26 March 2026  04:21:45 +0000 (0:00:02.810)       0:00:09.943 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nThursday 26 March 2026  04:21:47 +0000 (0:00:01.954)       0:00:11.898 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nThursday 26 March 2026  04:21:49 +0000 (0:00:01.980)       0:00:13.878 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nThursday 26 March 2026  04:21:51 +0000 (0:00:02.413)       0:00:16.291 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nThursday 26 March 2026  04:21:53 +0000 (0:00:01.167)       0:00:17.459 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nThursday 26 March 2026  04:21:54 +0000 (0:00:01.140)       0:00:18.600 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nThursday 26 March 2026  04:21:56 +0000 (0:00:01.916)       0:00:20.516 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nThursday 26 March 2026  04:21:59 +0000 (0:00:03.629)       0:00:24.146 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:22:02 +0000 (0:00:02.615)       0:00:26.761 ******** \n=============================================================================== \nosism.services.frr : Set sysctl parameters ------------------------------ 3.63s\nosism.services.frr : Install frr package -------------------------------- 2.81s\nosism.services.frr : Manage frr service --------------------------------- 2.62s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 2.41s\nosism.services.frr : Include distribution specific install tasks -------- 2.11s\nosism.services.frr : Pin frr package version ---------------------------- 2.06s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 1.98s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.95s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.92s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 1.17s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 1.14s\n+ osism apply kubernetes\n\u001b[32m2026-03-26 04:22:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe2e5425-917f-42a5-93e8-ab4ed1c92939 (kubernetes) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:22:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fe2e5425-917f-42a5-93e8-ab4ed1c92939 (kubernetes) has been started and output is visible here.\u001b[0m\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nThursday 26 March 2026  04:22:11 +0000 (0:00:01.796)       0:00:01.796 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nThursday 26 March 2026  04:22:16 +0000 (0:00:05.407)       0:00:07.204 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nThursday 26 March 2026  04:22:18 +0000 (0:00:01.878)       0:00:09.082 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nThursday 26 March 2026  04:22:20 +0000 (0:00:01.950)       0:00:11.033 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nThursday 26 March 2026  04:22:23 +0000 (0:00:03.154)       0:00:14.187 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nThursday 26 March 2026  04:22:26 +0000 (0:00:02.308)       0:00:16.496 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nThursday 26 March 2026  04:22:28 +0000 (0:00:02.093)       0:00:18.590 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nThursday 26 March 2026  04:22:30 +0000 (0:00:02.232)       0:00:20.822 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nThursday 26 March 2026  04:22:32 +0000 (0:00:01.722)       0:00:22.544 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nThursday 26 March 2026  04:22:34 +0000 (0:00:01.977)       0:00:24.522 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nThursday 26 March 2026  04:22:36 +0000 (0:00:02.055)       0:00:26.577 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nThursday 26 March 2026  04:22:39 +0000 (0:00:03.010)       0:00:29.588 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nThursday 26 March 2026  04:22:41 +0000 (0:00:02.691)       0:00:32.280 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nThursday 26 March 2026  04:22:43 +0000 (0:00:01.934)       0:00:34.214 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nThursday 26 March 2026  04:22:46 +0000 (0:00:02.286)       0:00:36.500 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nThursday 26 March 2026  04:22:47 +0000 (0:00:01.698)       0:00:38.199 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nThursday 26 March 2026  04:22:49 +0000 (0:00:02.052)       0:00:40.251 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nThursday 26 March 2026  04:22:51 +0000 (0:00:01.867)       0:00:42.119 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nThursday 26 March 2026  04:22:54 +0000 (0:00:02.859)       0:00:44.979 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nThursday 26 March 2026  04:22:56 +0000 (0:00:01.767)       0:00:46.747 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nThursday 26 March 2026  04:22:58 +0000 (0:00:02.080)       0:00:48.828 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nThursday 26 March 2026  04:23:00 +0000 (0:00:02.134)       0:00:50.962 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nThursday 26 March 2026  04:23:02 +0000 (0:00:01.964)       0:00:52.926 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nThursday 26 March 2026  04:23:03 +0000 (0:00:01.363)       0:00:54.290 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nThursday 26 March 2026  04:23:05 +0000 (0:00:01.705)       0:00:55.995 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nThursday 26 March 2026  04:23:07 +0000 (0:00:02.200)       0:00:58.196 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nThursday 26 March 2026  04:23:09 +0000 (0:00:01.908)       0:01:00.104 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nThursday 26 March 2026  04:23:12 +0000 (0:00:02.526)       0:01:02.631 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nThursday 26 March 2026  04:23:13 +0000 (0:00:01.669)       0:01:04.300 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nThursday 26 March 2026  04:23:15 +0000 (0:00:01.858)       0:01:06.159 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nThursday 26 March 2026  04:23:18 +0000 (0:00:02.518)       0:01:08.677 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nThursday 26 March 2026  04:23:19 +0000 (0:00:01.368)       0:01:10.046 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nThursday 26 March 2026  04:23:21 +0000 (0:00:01.699)       0:01:11.745 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nThursday 26 March 2026  04:23:23 +0000 (0:00:02.229)       0:01:13.975 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nThursday 26 March 2026  04:23:25 +0000 (0:00:01.820)       0:01:15.796 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nThursday 26 March 2026  04:23:26 +0000 (0:00:01.411)       0:01:17.208 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nThursday 26 March 2026  04:23:50 +0000 (0:00:23.365)       0:01:40.573 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nThursday 26 March 2026  04:23:51 +0000 (0:00:01.362)       0:01:41.936 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nThursday 26 March 2026  04:23:53 +0000 (0:00:02.075)       0:01:44.011 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nThursday 26 March 2026  04:23:55 +0000 (0:00:02.314)       0:01:46.326 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nThursday 26 March 2026  04:24:23 +0000 (0:00:27.061)       0:02:13.387 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nThursday 26 March 2026  04:24:24 +0000 (0:00:01.800)       0:02:15.187 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nThursday 26 March 2026  04:24:26 +0000 (0:00:01.708)       0:02:16.895 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nThursday 26 March 2026  04:24:28 +0000 (0:00:01.997)       0:02:18.893 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nThursday 26 March 2026  04:24:30 +0000 (0:00:01.744)       0:02:20.638 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nThursday 26 March 2026  04:24:31 +0000 (0:00:01.384)       0:02:22.023 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nThursday 26 March 2026  04:24:33 +0000 (0:00:01.654)       0:02:23.677 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nThursday 26 March 2026  04:24:35 +0000 (0:00:01.978)       0:02:25.655 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nThursday 26 March 2026  04:24:37 +0000 (0:00:01.865)       0:02:27.520 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nThursday 26 March 2026  04:24:39 +0000 (0:00:02.004)       0:02:29.525 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nThursday 26 March 2026  04:24:40 +0000 (0:00:01.370)       0:02:30.896 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nThursday 26 March 2026  04:24:41 +0000 (0:00:01.317)       0:02:32.214 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nThursday 26 March 2026  04:24:43 +0000 (0:00:01.658)       0:02:33.872 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nThursday 26 March 2026  04:24:45 +0000 (0:00:01.650)       0:02:35.523 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nThursday 26 March 2026  04:24:49 +0000 (0:00:04.287)       0:02:39.810 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nThursday 26 March 2026  04:24:50 +0000 (0:00:01.359)       0:02:41.170 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nThursday 26 March 2026  04:24:52 +0000 (0:00:01.639)       0:02:42.809 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nThursday 26 March 2026  04:24:54 +0000 (0:00:01.553)       0:02:44.363 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nThursday 26 March 2026  04:24:55 +0000 (0:00:01.704)       0:02:46.067 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nThursday 26 March 2026  04:24:57 +0000 (0:00:01.488)       0:02:47.556 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nThursday 26 March 2026  04:24:58 +0000 (0:00:01.376)       0:02:48.932 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nThursday 26 March 2026  04:25:00 +0000 (0:00:01.437)       0:02:50.370 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nThursday 26 March 2026  04:25:01 +0000 (0:00:01.714)       0:02:52.085 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nThursday 26 March 2026  04:25:03 +0000 (0:00:02.180)       0:02:54.265 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nThursday 26 March 2026  04:25:06 +0000 (0:00:02.250)       0:02:56.516 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nThursday 26 March 2026  04:25:14 +0000 (0:00:07.927)       0:03:04.443 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nThursday 26 March 2026  04:25:16 +0000 (0:00:02.221)       0:03:06.664 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nThursday 26 March 2026  04:25:17 +0000 (0:00:01.434)       0:03:08.099 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nThursday 26 March 2026  04:25:19 +0000 (0:00:01.540)       0:03:09.640 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nThursday 26 March 2026  04:25:21 +0000 (0:00:01.925)       0:03:11.565 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nThursday 26 March 2026  04:25:22 +0000 (0:00:01.537)       0:03:13.103 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nThursday 26 March 2026  04:25:25 +0000 (0:00:02.937)       0:03:16.040 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nThursday 26 March 2026  04:25:27 +0000 (0:00:01.787)       0:03:17.828 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nThursday 26 March 2026  04:25:28 +0000 (0:00:01.448)       0:03:19.277 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nThursday 26 March 2026  04:25:30 +0000 (0:00:01.867)       0:03:21.144 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nThursday 26 March 2026  04:25:31 +0000 (0:00:01.154)       0:03:22.299 ******** \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nThursday 26 March 2026  04:25:33 +0000 (0:00:01.486)       0:03:23.785 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nThursday 26 March 2026  04:25:35 +0000 (0:00:01.834)       0:03:25.620 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nThursday 26 March 2026  04:25:37 +0000 (0:00:02.573)       0:03:28.194 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nThursday 26 March 2026  04:25:39 +0000 (0:00:01.514)       0:03:29.709 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nThursday 26 March 2026  04:25:40 +0000 (0:00:01.433)       0:03:31.143 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nThursday 26 March 2026  04:25:42 +0000 (0:00:01.654)       0:03:32.798 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nThursday 26 March 2026  04:25:44 +0000 (0:00:02.418)       0:03:35.217 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nThursday 26 March 2026  04:25:46 +0000 (0:00:01.700)       0:03:36.917 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nThursday 26 March 2026  04:25:47 +0000 (0:00:01.410)       0:03:38.328 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nThursday 26 March 2026  04:25:49 +0000 (0:00:01.715)       0:03:40.044 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nThursday 26 March 2026  04:25:51 +0000 (0:00:01.847)       0:03:41.892 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nThursday 26 March 2026  04:25:53 +0000 (0:00:01.905)       0:03:43.797 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nThursday 26 March 2026  04:25:55 +0000 (0:00:01.865)       0:03:45.663 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nThursday 26 March 2026  04:25:56 +0000 (0:00:01.151)       0:03:46.815 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nThursday 26 March 2026  04:25:58 +0000 (0:00:01.972)       0:03:48.788 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nThursday 26 March 2026  04:26:00 +0000 (0:00:02.341)       0:03:51.130 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nThursday 26 March 2026  04:26:01 +0000 (0:00:01.175)       0:03:52.306 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nThursday 26 March 2026  04:26:03 +0000 (0:00:01.170)       0:03:53.476 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"Installed Cilium version: 1.18.2, Target Cilium version: v1.18.2, Update needed: False\\n\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nThursday 26 March 2026  04:26:04 +0000 (0:00:01.149)       0:03:54.626 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nThursday 26 March 2026  04:26:05 +0000 (0:00:01.155)       0:03:55.782 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nThursday 26 March 2026  04:26:10 +0000 (0:00:05.467)       0:04:01.250 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nThursday 26 March 2026  04:26:13 +0000 (0:00:02.383)       0:04:03.634 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nThursday 26 March 2026  04:26:15 +0000 (0:00:02.620)       0:04:06.254 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nThursday 26 March 2026  04:26:19 +0000 (0:00:04.090)       0:04:10.345 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nThursday 26 March 2026  04:26:21 +0000 (0:00:01.145)       0:04:11.491 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nThursday 26 March 2026  04:26:24 +0000 (0:00:02.908)       0:04:14.400 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nThursday 26 March 2026  04:26:25 +0000 (0:00:01.461)       0:04:15.861 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nThursday 26 March 2026  04:26:27 +0000 (0:00:02.085)       0:04:17.946 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nThursday 26 March 2026  04:26:28 +0000 (0:00:01.326)       0:04:19.273 ******** \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nThursday 26 March 2026  04:26:30 +0000 (0:00:01.561)       0:04:20.835 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nThursday 26 March 2026  04:26:35 +0000 (0:00:05.003)       0:04:25.839 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nThursday 26 March 2026  04:26:37 +0000 (0:00:01.875)       0:04:27.715 ******** \n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nThursday 26 March 2026  04:26:46 +0000 (0:00:09.159)       0:04:36.874 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nThursday 26 March 2026  04:26:48 +0000 (0:00:01.861)       0:04:38.736 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=53  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:26:50 +0000 (0:00:02.479)       0:04:41.215 ******** \n=============================================================================== \nk3s_server : Enable and check K3s service ------------------------------ 27.06s\nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 23.37s\nManage labels ----------------------------------------------------------- 9.16s\nk3s_agent : Manage k3s service ------------------------------------------ 7.93s\nk3s_server_post : Wait for Cilium resources ----------------------------- 5.47s\nk3s_prereq : Validating arguments against arg spec 'main' - Prerequisites --- 5.41s\nk9s : Install k9s packages ---------------------------------------------- 5.00s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 4.29s\nk3s_server_post : Apply BGP manifests ----------------------------------- 4.09s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 3.15s\nk3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries --- 3.01s\nMake kubeconfig available for use inside the manager service ------------ 2.94s\nk3s_server_post : Test for BGP config resources ------------------------- 2.91s\nk3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 2.86s\nk3s_download : Download k3s binary x64 ---------------------------------- 2.69s\nk3s_server_post : Copy BGP manifests to first master -------------------- 2.62s\nkubectl : Install apt-transport-https package --------------------------- 2.57s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.53s\nk3s_server : Copy vip manifest to first master -------------------------- 2.52s\nManage taints ----------------------------------------------------------- 2.48s\n+ [[ false == \\f\\a\\l\\s\\e ]]\n+ sh -c /opt/configuration/scripts/upgrade/200-infrastructure.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ osism apply openstackclient\n\u001b[32m2026-03-26 04:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e1a66714-e3a4-43c7-84c4-5634d0d755e8 (openstackclient) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e1a66714-e3a4-43c7-84c4-5634d0d755e8 (openstackclient) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x7b96ee349be0>): Expecting value: line 2 column 1 (char 1)\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x7b96ee349be0>): 'NoneType' object is not subscriptable\u001b[0m\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nThursday 26 March 2026  04:27:10 +0000 (0:00:01.545)       0:00:01.545 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nThursday 26 March 2026  04:27:10 +0000 (0:00:00.865)       0:00:02.410 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nThursday 26 March 2026  04:27:12 +0000 (0:00:01.489)       0:00:03.900 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nThursday 26 March 2026  04:27:13 +0000 (0:00:01.268)       0:00:05.168 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nThursday 26 March 2026  04:27:14 +0000 (0:00:01.125)       0:00:06.293 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nThursday 26 March 2026  04:27:15 +0000 (0:00:00.931)       0:00:07.225 ******** \n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m\n\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x7b96ee349be0>): 'NoneType' object is not subscriptable\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nThursday 26 March 2026  04:27:16 +0000 (0:00:00.713)       0:00:07.938 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nThursday 26 March 2026  04:27:26 +0000 (0:00:10.507)       0:00:18.446 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nThursday 26 March 2026  04:27:28 +0000 (0:00:01.309)       0:00:19.755 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nThursday 26 March 2026  04:27:28 +0000 (0:00:00.616)       0:00:20.371 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:27:30 +0000 (0:00:01.156)       0:00:21.528 ******** \n=============================================================================== \nosism.services.openstackclient : Restart openstackclient service ------- 10.51s\nosism.services.openstackclient : Create required directories ------------ 1.49s\nosism.services.openstackclient : Ensure that all containers are up ------ 1.31s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 1.27s\nosism.services.openstackclient : Copy bash completion script ------------ 1.16s\nosism.services.openstackclient : Manage openstackclient service --------- 1.13s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 0.93s\nosism.services.openstackclient : Include tasks -------------------------- 0.87s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 0.71s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.62s\n+ osism apply -a upgrade common\n\u001b[32m2026-03-26 04:27:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d9da0468-b2fe-479e-8776-94e7f6347ff6 (common) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:27:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d9da0468-b2fe-479e-8776-94e7f6347ff6 (common) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  04:27:39 +0000 (0:00:02.502)       0:00:02.502 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/upgrade.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nThursday 26 March 2026  04:27:43 +0000 (0:00:03.336)       0:00:05.838 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  04:27:46 +0000 (0:00:03.585)       0:00:09.424 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nThursday 26 March 2026  04:27:49 +0000 (0:00:02.808)       0:00:12.232 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:27:53 +0000 (0:00:04.313)       0:00:16.546 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nThursday 26 March 2026  04:27:56 +0000 (0:00:02.875)       0:00:19.421 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensure /var/log/journal exists on EL10 systems] *****************\nThursday 26 March 2026  04:27:59 +0000 (0:00:03.287)       0:00:22.708 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nThursday 26 March 2026  04:28:02 +0000 (0:00:02.349)       0:00:25.058 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nThursday 26 March 2026  04:28:04 +0000 (0:00:02.157)       0:00:27.215 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over kolla.target] **************************************\nThursday 26 March 2026  04:28:06 +0000 (0:00:02.257)       0:00:29.473 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nThursday 26 March 2026  04:28:09 +0000 (0:00:02.933)       0:00:32.406 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nThursday 26 March 2026  04:28:14 +0000 (0:00:04.925)       0:00:37.331 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nThursday 26 March 2026  04:28:16 +0000 (0:00:02.395)       0:00:39.727 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nThursday 26 March 2026  04:28:18 +0000 (0:00:01.857)       0:00:41.585 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nThursday 26 March 2026  04:28:20 +0000 (0:00:01.853)       0:00:43.438 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nThursday 26 March 2026  04:28:22 +0000 (0:00:01.835)       0:00:45.273 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nThursday 26 March 2026  04:28:26 +0000 (0:00:04.267)       0:00:49.541 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nThursday 26 March 2026  04:28:29 +0000 (0:00:03.102)       0:00:52.643 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nThursday 26 March 2026  04:28:32 +0000 (0:00:02.866)       0:00:55.509 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nThursday 26 March 2026  04:28:35 +0000 (0:00:02.813)       0:00:58.323 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nThursday 26 March 2026  04:28:38 +0000 (0:00:03.118)       0:01:01.441 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [service-check-containers : common | Check containers] ********************\nThursday 26 March 2026  04:28:41 +0000 (0:00:03.309)       0:01:04.751 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : common | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:28:46 +0000 (0:00:04.481)       0:01:09.233 ******** \n\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:28:48 +0000 (0:00:02.103)       0:01:11.337 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:51 +0000 (0:00:03.054)       0:01:14.392 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:52 +0000 (0:00:00.446)       0:01:14.838 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:52 +0000 (0:00:00.441)       0:01:15.280 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:52 +0000 (0:00:00.442)       0:01:15.722 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:53 +0000 (0:00:00.461)       0:01:16.183 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:54 +0000 (0:00:00.715)       0:01:16.899 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:28:54 +0000 (0:00:00.443)       0:01:17.342 ******** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nThursday 26 March 2026  04:28:55 +0000 (0:00:00.829)       0:01:18.172 ******** \n\u001b[0;31mfatal: [testbed-manager]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[0;31mfatal: [testbed-node-3]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[0;31mfatal: [testbed-node-4]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[32m2026-03-26 04:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62611976-a838-4c25-9ed0-48c8dae1f80c (common) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62611976-a838-4c25-9ed0-48c8dae1f80c (common) has been started and output is visible here.\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\u001b[0;31mfatal: [testbed-node-5]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;31mtestbed-manager\u001b[0m            : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-0\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-3\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-4\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;31mtestbed-node-5\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:29:01 +0000 (0:00:06.225)       0:01:24.397 ******** \n=============================================================================== \ncommon : Restart fluentd container -------------------------------------- 6.23s\ncommon : Copying over config.json files for services -------------------- 4.93s\nservice-check-containers : common | Check containers -------------------- 4.48s\nservice-cert-copy : common | Copying over extra CA certificates --------- 4.31s\ncommon : Copying over fluentd.conf -------------------------------------- 4.27s\ncommon : Flush handlers ------------------------------------------------- 3.78s\ncommon : Ensuring config directories exist ------------------------------ 3.59s\ncommon : include_tasks -------------------------------------------------- 3.34s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 3.31s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 3.29s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 3.12s\ncommon : Copying over cron logrotate config file ------------------------ 3.10s\nservice-check-containers : Include tasks -------------------------------- 3.05s\ncommon : Copying over kolla.target -------------------------------------- 2.93s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 2.88s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.87s\ncommon : Ensuring config directories have correct owner and permission --- 2.81s\ncommon : include_tasks -------------------------------------------------- 2.81s\ncommon : Find custom fluentd input config files ------------------------- 2.40s\ncommon : Ensure /var/log/journal exists on EL10 systems ----------------- 2.35s\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x774c1743cad0>): Expecting value: line 2 column 1 (char 1)\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x774c1743cad0>): 'NoneType' object is not subscriptable\u001b[0m\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  04:29:07 +0000 (0:00:01.579)       0:00:01.581 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/upgrade.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nThursday 26 March 2026  04:29:10 +0000 (0:00:02.455)       0:00:04.036 ******** \n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nThursday 26 March 2026  04:29:12 +0000 (0:00:02.257)       0:00:06.294 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nThursday 26 March 2026  04:29:14 +0000 (0:00:02.087)       0:00:08.381 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:29:18 +0000 (0:00:03.314)       0:00:11.696 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nThursday 26 March 2026  04:29:19 +0000 (0:00:01.538)       0:00:13.234 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensure /var/log/journal exists on EL10 systems] *****************\nThursday 26 March 2026  04:29:21 +0000 (0:00:02.256)       0:00:15.490 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nThursday 26 March 2026  04:29:22 +0000 (0:00:01.071)       0:00:16.562 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nThursday 26 March 2026  04:29:23 +0000 (0:00:00.929)       0:00:17.492 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over kolla.target] **************************************\nThursday 26 March 2026  04:29:24 +0000 (0:00:00.789)       0:00:18.281 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nThursday 26 March 2026  04:29:26 +0000 (0:00:01.923)       0:00:20.205 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nThursday 26 March 2026  04:29:30 +0000 (0:00:03.670)       0:00:23.875 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nThursday 26 March 2026  04:29:31 +0000 (0:00:01.454)       0:00:25.330 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nThursday 26 March 2026  04:29:32 +0000 (0:00:00.897)       0:00:26.228 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nThursday 26 March 2026  04:29:33 +0000 (0:00:00.949)       0:00:27.177 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nThursday 26 March 2026  04:29:34 +0000 (0:00:00.935)       0:00:28.112 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nThursday 26 March 2026  04:29:37 +0000 (0:00:02.815)       0:00:30.927 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nThursday 26 March 2026  04:29:39 +0000 (0:00:02.467)       0:00:33.395 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nThursday 26 March 2026  04:29:41 +0000 (0:00:01.984)       0:00:35.380 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nThursday 26 March 2026  04:29:43 +0000 (0:00:02.110)       0:00:37.490 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nThursday 26 March 2026  04:29:46 +0000 (0:00:02.170)       0:00:39.661 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [service-check-containers : common | Check containers] ********************\nThursday 26 March 2026  04:29:48 +0000 (0:00:02.193)       0:00:41.854 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : common | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:29:51 +0000 (0:00:03.167)       0:00:45.021 ******** \n\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:29:52 +0000 (0:00:01.045)       0:00:46.067 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m\n\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x774c1743cad0>): 'NoneType' object is not subscriptable\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:54 +0000 (0:00:02.275)       0:00:48.342 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:54 +0000 (0:00:00.090)       0:00:48.433 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:54 +0000 (0:00:00.075)       0:00:48.508 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:54 +0000 (0:00:00.072)       0:00:48.581 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:55 +0000 (0:00:00.072)       0:00:48.653 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:55 +0000 (0:00:00.369)       0:00:49.022 ******** \n\nTASK [common : Flush handlers] *************************************************\nThursday 26 March 2026  04:29:55 +0000 (0:00:00.080)       0:00:49.103 ******** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nThursday 26 March 2026  04:29:55 +0000 (0:00:00.109)       0:00:49.212 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nThursday 26 March 2026  04:30:33 +0000 (0:00:37.548)       0:01:26.760 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nThursday 26 March 2026  04:31:12 +0000 (0:00:39.056)       0:02:05.817 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nThursday 26 March 2026  04:31:14 +0000 (0:00:01.973)       0:02:07.790 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:31:26 +0000 (0:00:11.972)       0:02:19.762 ******** \n=============================================================================== \ncommon : Restart kolla-toolbox container ------------------------------- 39.06s\ncommon : Restart fluentd container ------------------------------------- 37.55s\ncommon : Restart cron container ---------------------------------------- 11.97s\ncommon : Copying over config.json files for services -------------------- 3.67s\nservice-cert-copy : common | Copying over extra CA certificates --------- 3.31s\nservice-check-containers : common | Check containers -------------------- 3.17s\ncommon : Copying over fluentd.conf -------------------------------------- 2.82s\ncommon : Copying over cron logrotate config file ------------------------ 2.47s\ncommon : include_tasks -------------------------------------------------- 2.46s\nservice-check-containers : Include tasks -------------------------------- 2.28s\ncommon : Ensuring config directories exist ------------------------------ 2.26s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 2.26s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.19s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.17s\ncommon : Ensuring config directories have correct owner and permission --- 2.11s\ncommon : include_tasks -------------------------------------------------- 2.09s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 1.99s\ncommon : Initializing toolbox container using normal user --------------- 1.97s\ncommon : Copying over kolla.target -------------------------------------- 1.92s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 1.54s\n+ osism apply -a upgrade loadbalancer\n\u001b[32m2026-03-26 04:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5147d542-5a4e-40bb-a5dc-01c2c9e8a857 (loadbalancer) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5147d542-5a4e-40bb-a5dc-01c2c9e8a857 (loadbalancer) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:31:36 +0000 (0:00:02.439)       0:00:02.439 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:31:38 +0000 (0:00:02.336)       0:00:04.775 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nThursday 26 March 2026  04:31:42 +0000 (0:00:03.757)       0:00:08.533 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Stop and remove containers for haproxy exporter containers] ***\nThursday 26 March 2026  04:31:44 +0000 (0:00:02.088)       0:00:10.621 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing config for haproxy exporter] *********************\nThursday 26 March 2026  04:31:46 +0000 (0:00:02.060)       0:00:12.682 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nThursday 26 March 2026  04:31:48 +0000 (0:00:02.137)       0:00:14.820 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nThursday 26 March 2026  04:31:50 +0000 (0:00:01.839)       0:00:16.660 ******** \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nThursday 26 March 2026  04:31:52 +0000 (0:00:01.921)       0:00:18.581 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nThursday 26 March 2026  04:31:53 +0000 (0:00:01.697)       0:00:20.279 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nThursday 26 March 2026  04:31:58 +0000 (0:00:04.044)       0:00:24.323 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nThursday 26 March 2026  04:32:00 +0000 (0:00:02.183)       0:00:26.507 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nThursday 26 March 2026  04:32:02 +0000 (0:00:02.350)       0:00:28.858 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nThursday 26 March 2026  04:32:04 +0000 (0:00:01.899)       0:00:30.758 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nThursday 26 March 2026  04:32:07 +0000 (0:00:02.672)       0:00:33.430 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nThursday 26 March 2026  04:32:09 +0000 (0:00:01.982)       0:00:35.412 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nThursday 26 March 2026  04:32:11 +0000 (0:00:02.727)       0:00:38.141 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nThursday 26 March 2026  04:32:14 +0000 (0:00:02.315)       0:00:40.456 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nThursday 26 March 2026  04:32:16 +0000 (0:00:02.341)       0:00:42.798 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nThursday 26 March 2026  04:32:18 +0000 (0:00:01.712)       0:00:44.511 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nThursday 26 March 2026  04:32:23 +0000 (0:00:05.084)       0:00:49.596 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nThursday 26 March 2026  04:32:28 +0000 (0:00:04.771)       0:00:54.367 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nThursday 26 March 2026  04:32:30 +0000 (0:00:02.726)       0:00:57.094 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nThursday 26 March 2026  04:32:35 +0000 (0:00:04.402)       0:01:01.496 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nThursday 26 March 2026  04:32:37 +0000 (0:00:01.896)       0:01:03.393 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nThursday 26 March 2026  04:32:40 +0000 (0:00:02.967)       0:01:06.361 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nThursday 26 March 2026  04:32:42 +0000 (0:00:02.798)       0:01:09.159 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nThursday 26 March 2026  04:32:44 +0000 (0:00:01.863)       0:01:11.023 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nThursday 26 March 2026  04:32:47 +0000 (0:00:02.690)       0:01:13.713 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql-cert.pem] ***************************\nThursday 26 March 2026  04:32:49 +0000 (0:00:02.590)       0:01:16.303 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql-key.pem] ****************************\nThursday 26 March 2026  04:32:51 +0000 (0:00:01.409)       0:01:17.713 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nThursday 26 March 2026  04:32:53 +0000 (0:00:01.933)       0:01:19.647 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:32:57 +0000 (0:00:04.123)       0:01:23.771 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nThursday 26 March 2026  04:32:59 +0000 (0:00:01.634)       0:01:25.405 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nThursday 26 March 2026  04:33:01 +0000 (0:00:02.100)       0:01:27.505 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nThursday 26 March 2026  04:33:03 +0000 (0:00:02.576)       0:01:30.082 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nThursday 26 March 2026  04:33:06 +0000 (0:00:02.495)       0:01:32.577 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Check containers] **************\nThursday 26 March 2026  04:33:08 +0000 (0:00:02.658)       0:01:35.236 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:33:13 +0000 (0:00:04.113)       0:01:39.350 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:33:14 +0000 (0:00:01.395)       0:01:40.745 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nThursday 26 March 2026  04:33:16 +0000 (0:00:02.265)       0:01:43.010 ******** \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nThursday 26 March 2026  04:33:18 +0000 (0:00:02.022)       0:01:45.032 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nThursday 26 March 2026  04:33:23 +0000 (0:00:04.748)       0:01:49.781 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nThursday 26 March 2026  04:33:25 +0000 (0:00:01.719)       0:01:51.501 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nThursday 26 March 2026  04:33:27 +0000 (0:00:02.425)       0:01:53.927 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nThursday 26 March 2026  04:33:29 +0000 (0:00:02.386)       0:01:56.313 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nThursday 26 March 2026  04:33:32 +0000 (0:00:02.961)       0:01:59.275 ******** \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nThursday 26 March 2026  04:33:34 +0000 (0:00:01.605)       0:02:00.881 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nThursday 26 March 2026  04:33:39 +0000 (0:00:04.523)       0:02:05.404 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nThursday 26 March 2026  04:33:40 +0000 (0:00:01.626)       0:02:07.030 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nThursday 26 March 2026  04:33:42 +0000 (0:00:01.806)       0:02:08.837 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nThursday 26 March 2026  04:33:44 +0000 (0:00:02.341)       0:02:11.178 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nThursday 26 March 2026  04:33:47 +0000 (0:00:02.844)       0:02:14.023 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nThursday 26 March 2026  04:33:49 +0000 (0:00:01.352)       0:02:15.375 ******** \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nThursday 26 March 2026  04:33:50 +0000 (0:00:01.728)       0:02:17.104 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nThursday 26 March 2026  04:33:54 +0000 (0:00:03.538)       0:02:20.643 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nThursday 26 March 2026  04:33:56 +0000 (0:00:02.633)       0:02:23.276 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nThursday 26 March 2026  04:33:59 +0000 (0:00:02.867)       0:02:26.143 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nThursday 26 March 2026  04:34:01 +0000 (0:00:01.591)       0:02:27.735 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nThursday 26 March 2026  04:34:03 +0000 (0:00:02.474)       0:02:30.209 ******** \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nThursday 26 March 2026  04:34:05 +0000 (0:00:01.805)       0:02:32.015 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nThursday 26 March 2026  04:34:10 +0000 (0:00:04.835)       0:02:36.850 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nThursday 26 March 2026  04:34:12 +0000 (0:00:01.941)       0:02:38.792 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nThursday 26 March 2026  04:34:14 +0000 (0:00:02.014)       0:02:40.806 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nThursday 26 March 2026  04:34:16 +0000 (0:00:02.237)       0:02:43.043 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nThursday 26 March 2026  04:34:19 +0000 (0:00:02.849)       0:02:45.893 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nThursday 26 March 2026  04:34:21 +0000 (0:00:01.567)       0:02:47.460 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nThursday 26 March 2026  04:34:22 +0000 (0:00:01.398)       0:02:48.859 ******** \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nThursday 26 March 2026  04:34:24 +0000 (0:00:01.783)       0:02:50.643 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nThursday 26 March 2026  04:34:29 +0000 (0:00:04.864)       0:02:55.508 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nThursday 26 March 2026  04:34:31 +0000 (0:00:01.914)       0:02:57.422 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nThursday 26 March 2026  04:34:33 +0000 (0:00:02.174)       0:02:59.597 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nThursday 26 March 2026  04:34:35 +0000 (0:00:02.304)       0:03:01.901 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nThursday 26 March 2026  04:34:38 +0000 (0:00:02.923)       0:03:04.825 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nThursday 26 March 2026  04:34:39 +0000 (0:00:01.332)       0:03:06.157 ******** \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nThursday 26 March 2026  04:34:41 +0000 (0:00:01.819)       0:03:07.977 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nThursday 26 March 2026  04:34:47 +0000 (0:00:05.708)       0:03:13.686 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nThursday 26 March 2026  04:34:51 +0000 (0:00:04.576)       0:03:18.262 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nThursday 26 March 2026  04:34:56 +0000 (0:00:04.653)       0:03:22.916 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nThursday 26 March 2026  04:34:58 +0000 (0:00:02.295)       0:03:25.211 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nThursday 26 March 2026  04:35:01 +0000 (0:00:02.964)       0:03:28.175 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nThursday 26 March 2026  04:35:03 +0000 (0:00:01.641)       0:03:29.817 ******** \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nThursday 26 March 2026  04:35:05 +0000 (0:00:01.735)       0:03:31.553 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nThursday 26 March 2026  04:35:09 +0000 (0:00:04.170)       0:03:35.723 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nThursday 26 March 2026  04:35:10 +0000 (0:00:01.512)       0:03:37.235 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nThursday 26 March 2026  04:35:12 +0000 (0:00:01.381)       0:03:38.616 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nThursday 26 March 2026  04:35:14 +0000 (0:00:02.275)       0:03:40.892 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nThursday 26 March 2026  04:35:17 +0000 (0:00:02.862)       0:03:43.755 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nThursday 26 March 2026  04:35:18 +0000 (0:00:01.425)       0:03:45.180 ******** \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nThursday 26 March 2026  04:35:20 +0000 (0:00:01.779)       0:03:46.960 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nThursday 26 March 2026  04:35:25 +0000 (0:00:04.915)       0:03:51.875 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nThursday 26 March 2026  04:35:27 +0000 (0:00:01.701)       0:03:53.577 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nThursday 26 March 2026  04:35:29 +0000 (0:00:02.029)       0:03:55.607 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nThursday 26 March 2026  04:35:31 +0000 (0:00:02.253)       0:03:57.860 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nThursday 26 March 2026  04:35:34 +0000 (0:00:02.893)       0:04:00.754 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nThursday 26 March 2026  04:35:35 +0000 (0:00:01.395)       0:04:02.150 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nThursday 26 March 2026  04:35:37 +0000 (0:00:01.364)       0:04:03.514 ******** \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nThursday 26 March 2026  04:35:39 +0000 (0:00:02.025)       0:04:05.540 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nThursday 26 March 2026  04:35:44 +0000 (0:00:04.854)       0:04:10.395 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nThursday 26 March 2026  04:35:46 +0000 (0:00:02.048)       0:04:12.444 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nThursday 26 March 2026  04:35:47 +0000 (0:00:01.707)       0:04:14.151 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nThursday 26 March 2026  04:35:50 +0000 (0:00:02.283)       0:04:16.434 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nThursday 26 March 2026  04:35:53 +0000 (0:00:03.304)       0:04:19.739 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nThursday 26 March 2026  04:35:54 +0000 (0:00:01.383)       0:04:21.123 ******** \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nThursday 26 March 2026  04:35:56 +0000 (0:00:01.920)       0:04:23.044 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nThursday 26 March 2026  04:36:01 +0000 (0:00:04.911)       0:04:27.955 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nThursday 26 March 2026  04:36:03 +0000 (0:00:01.862)       0:04:29.817 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nThursday 26 March 2026  04:36:05 +0000 (0:00:02.006)       0:04:31.823 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nThursday 26 March 2026  04:36:07 +0000 (0:00:02.371)       0:04:34.195 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nThursday 26 March 2026  04:36:10 +0000 (0:00:02.925)       0:04:37.121 ******** \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nThursday 26 March 2026  04:36:12 +0000 (0:00:02.165)       0:04:39.286 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nThursday 26 March 2026  04:36:17 +0000 (0:00:04.923)       0:04:44.209 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nThursday 26 March 2026  04:36:19 +0000 (0:00:01.716)       0:04:45.926 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nThursday 26 March 2026  04:36:21 +0000 (0:00:02.003)       0:04:47.929 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nThursday 26 March 2026  04:36:23 +0000 (0:00:02.254)       0:04:50.183 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nThursday 26 March 2026  04:36:26 +0000 (0:00:02.974)       0:04:53.158 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nThursday 26 March 2026  04:36:29 +0000 (0:00:02.622)       0:04:55.780 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nThursday 26 March 2026  04:36:33 +0000 (0:00:03.966)       0:04:59.747 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nThursday 26 March 2026  04:36:36 +0000 (0:00:03.513)       0:05:03.260 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nThursday 26 March 2026  04:36:40 +0000 (0:00:03.738)       0:05:06.999 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nThursday 26 March 2026  04:36:44 +0000 (0:00:03.918)       0:05:10.918 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nThursday 26 March 2026  04:36:47 +0000 (0:00:02.974)       0:05:13.893 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nThursday 26 March 2026  04:36:50 +0000 (0:00:02.727)       0:05:16.621 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nThursday 26 March 2026  04:36:51 +0000 (0:00:01.402)       0:05:18.023 ******** \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nThursday 26 March 2026  04:36:53 +0000 (0:00:02.176)       0:05:20.199 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nThursday 26 March 2026  04:36:56 +0000 (0:00:02.535)       0:05:22.735 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nThursday 26 March 2026  04:36:58 +0000 (0:00:01.745)       0:05:24.480 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nThursday 26 March 2026  04:36:59 +0000 (0:00:01.420)       0:05:25.901 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nThursday 26 March 2026  04:37:01 +0000 (0:00:01.454)       0:05:27.355 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nThursday 26 March 2026  04:37:03 +0000 (0:00:02.362)       0:05:29.717 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nThursday 26 March 2026  04:37:05 +0000 (0:00:01.707)       0:05:31.425 ******** \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nThursday 26 March 2026  04:37:07 +0000 (0:00:02.122)       0:05:33.547 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nThursday 26 March 2026  04:37:13 +0000 (0:00:06.368)       0:05:39.916 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nThursday 26 March 2026  04:37:16 +0000 (0:00:02.651)       0:05:42.568 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nThursday 26 March 2026  04:37:19 +0000 (0:00:02.989)       0:05:45.557 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nThursday 26 March 2026  04:37:21 +0000 (0:00:02.255)       0:05:47.813 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nThursday 26 March 2026  04:37:24 +0000 (0:00:02.986)       0:05:50.800 ******** \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nThursday 26 March 2026  04:37:26 +0000 (0:00:02.470)       0:05:53.271 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nThursday 26 March 2026  04:37:31 +0000 (0:00:04.663)       0:05:57.934 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nThursday 26 March 2026  04:37:33 +0000 (0:00:01.726)       0:05:59.660 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nThursday 26 March 2026  04:37:35 +0000 (0:00:01.907)       0:06:01.568 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nThursday 26 March 2026  04:37:37 +0000 (0:00:02.377)       0:06:03.945 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nThursday 26 March 2026  04:37:40 +0000 (0:00:02.975)       0:06:06.921 ******** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nThursday 26 March 2026  04:37:42 +0000 (0:00:02.309)       0:06:09.231 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nThursday 26 March 2026  04:37:50 +0000 (0:00:07.268)       0:06:16.500 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nThursday 26 March 2026  04:37:52 +0000 (0:00:01.951)       0:06:18.451 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nThursday 26 March 2026  04:37:54 +0000 (0:00:02.577)       0:06:21.029 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nThursday 26 March 2026  04:37:57 +0000 (0:00:02.333)       0:06:23.362 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nThursday 26 March 2026  04:38:00 +0000 (0:00:03.004)       0:06:26.367 ******** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nThursday 26 March 2026  04:38:02 +0000 (0:00:02.853)       0:06:29.220 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nThursday 26 March 2026  04:38:04 +0000 (0:00:01.765)       0:06:30.986 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nThursday 26 March 2026  04:38:09 +0000 (0:00:04.834)       0:06:35.820 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nThursday 26 March 2026  04:38:11 +0000 (0:00:02.439)       0:06:38.260 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  04:38:15 +0000 (0:00:03.286)       0:06:41.547 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  04:38:19 +0000 (0:00:03.873)       0:06:45.420 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nThursday 26 March 2026  04:38:22 +0000 (0:00:03.493)       0:06:48.914 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nThursday 26 March 2026  04:38:24 +0000 (0:00:01.563)       0:06:50.478 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nThursday 26 March 2026  04:38:26 +0000 (0:00:02.273)       0:06:52.752 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nThursday 26 March 2026  04:38:29 +0000 (0:00:02.574)       0:06:55.326 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  04:38:31 +0000 (0:00:02.312)       0:06:57.639 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  04:38:35 +0000 (0:00:03.768)       0:07:01.407 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nThursday 26 March 2026  04:38:39 +0000 (0:00:03.955)       0:07:05.363 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nThursday 26 March 2026  04:38:41 +0000 (0:00:02.298)       0:07:07.662 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nThursday 26 March 2026  04:38:43 +0000 (0:00:02.421)       0:07:10.083 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nThursday 26 March 2026  04:38:46 +0000 (0:00:02.444)       0:07:12.527 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nThursday 26 March 2026  04:38:48 +0000 (0:00:02.499)       0:07:15.027 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nThursday 26 March 2026  04:38:52 +0000 (0:00:03.669)       0:07:18.697 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nThursday 26 March 2026  04:38:56 +0000 (0:00:04.374)       0:07:23.071 ******** \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nThursday 26 March 2026  04:38:59 +0000 (0:00:02.593)       0:07:25.664 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nThursday 26 March 2026  04:39:04 +0000 (0:00:04.908)       0:07:30.572 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nThursday 26 March 2026  04:39:06 +0000 (0:00:02.124)       0:07:32.697 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nThursday 26 March 2026  04:39:08 +0000 (0:00:02.116)       0:07:34.813 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nThursday 26 March 2026  04:39:10 +0000 (0:00:02.256)       0:07:37.070 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nThursday 26 March 2026  04:39:13 +0000 (0:00:03.090)       0:07:40.160 ******** \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nThursday 26 March 2026  04:39:16 +0000 (0:00:02.574)       0:07:42.734 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nThursday 26 March 2026  04:39:23 +0000 (0:00:07.544)       0:07:50.279 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nThursday 26 March 2026  04:39:26 +0000 (0:00:02.320)       0:07:52.600 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nThursday 26 March 2026  04:39:28 +0000 (0:00:01.845)       0:07:54.446 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nThursday 26 March 2026  04:39:29 +0000 (0:00:01.483)       0:07:55.929 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nThursday 26 March 2026  04:39:31 +0000 (0:00:02.323)       0:07:58.252 ******** \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nThursday 26 March 2026  04:39:34 +0000 (0:00:02.722)       0:08:00.974 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nThursday 26 March 2026  04:39:40 +0000 (0:00:06.034)       0:08:07.009 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nThursday 26 March 2026  04:39:43 +0000 (0:00:02.577)       0:08:09.587 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nThursday 26 March 2026  04:39:45 +0000 (0:00:01.831)       0:08:11.419 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nThursday 26 March 2026  04:39:47 +0000 (0:00:01.975)       0:08:13.394 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nThursday 26 March 2026  04:39:49 +0000 (0:00:02.205)       0:08:15.599 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nThursday 26 March 2026  04:39:51 +0000 (0:00:02.350)       0:08:17.950 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nThursday 26 March 2026  04:39:55 +0000 (0:00:03.955)       0:08:21.905 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nThursday 26 March 2026  04:39:57 +0000 (0:00:01.510)       0:08:23.416 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nThursday 26 March 2026  04:39:58 +0000 (0:00:01.471)       0:08:24.888 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nThursday 26 March 2026  04:40:00 +0000 (0:00:01.974)       0:08:26.862 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nThursday 26 March 2026  04:40:02 +0000 (0:00:02.285)       0:08:29.147 ******** \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nThursday 26 March 2026  04:40:05 +0000 (0:00:02.516)       0:08:31.663 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nThursday 26 March 2026  04:40:13 +0000 (0:00:07.717)       0:08:39.381 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nThursday 26 March 2026  04:40:14 +0000 (0:00:01.751)       0:08:41.132 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nThursday 26 March 2026  04:40:17 +0000 (0:00:02.189)       0:08:43.322 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nThursday 26 March 2026  04:40:19 +0000 (0:00:02.377)       0:08:45.700 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nThursday 26 March 2026  04:40:22 +0000 (0:00:03.317)       0:08:49.017 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nThursday 26 March 2026  04:40:24 +0000 (0:00:01.504)       0:08:50.522 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nThursday 26 March 2026  04:40:25 +0000 (0:00:01.460)       0:08:51.983 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nThursday 26 March 2026  04:40:27 +0000 (0:00:01.827)       0:08:53.811 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nThursday 26 March 2026  04:40:28 +0000 (0:00:01.355)       0:08:55.166 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : loadbalancer] *********************************************\nThursday 26 March 2026  04:40:30 +0000 (0:00:01.430)       0:08:56.597 ******** \n\u001b[0;36mincluded: loadbalancer for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Check containers] **************\nThursday 26 March 2026  04:40:33 +0000 (0:00:02.818)       0:08:59.416 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:40:37 +0000 (0:00:04.383)       0:09:03.799 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:40:38 +0000 (0:00:01.442)       0:09:05.242 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nThursday 26 March 2026  04:40:41 +0000 (0:00:02.937)       0:09:08.180 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nThursday 26 March 2026  04:40:43 +0000 (0:00:01.803)       0:09:09.984 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nThursday 26 March 2026  04:40:45 +0000 (0:00:01.398)       0:09:11.382 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nThursday 26 March 2026  04:40:52 +0000 (0:00:07.025)       0:09:18.407 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nThursday 26 March 2026  04:40:59 +0000 (0:00:07.440)       0:09:25.848 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nThursday 26 March 2026  04:41:06 +0000 (0:00:07.139)       0:09:32.988 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nThursday 26 March 2026  04:41:14 +0000 (0:00:07.613)       0:09:40.602 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nThursday 26 March 2026  04:41:18 +0000 (0:00:03.825)       0:09:44.428 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nThursday 26 March 2026  04:41:31 +0000 (0:00:13.028)       0:09:57.456 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nThursday 26 March 2026  04:41:35 +0000 (0:00:04.614)       0:10:02.070 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nThursday 26 March 2026  04:41:43 +0000 (0:00:07.753)       0:10:09.824 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nThursday 26 March 2026  04:41:50 +0000 (0:00:06.838)       0:10:16.663 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nThursday 26 March 2026  04:41:57 +0000 (0:00:06.796)       0:10:23.460 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nThursday 26 March 2026  04:42:04 +0000 (0:00:06.957)       0:10:30.417 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for master haproxy to start] **************\nThursday 26 March 2026  04:42:11 +0000 (0:00:07.033)       0:10:37.451 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nThursday 26 March 2026  04:42:14 +0000 (0:00:03.645)       0:10:41.097 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for master proxysql to start] *************\nThursday 26 March 2026  04:42:27 +0000 (0:00:12.640)       0:10:53.737 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nThursday 26 March 2026  04:42:31 +0000 (0:00:03.624)       0:10:57.362 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nThursday 26 March 2026  04:42:37 +0000 (0:00:06.740)       0:11:04.102 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nThursday 26 March 2026  04:42:39 +0000 (0:00:01.974)       0:11:06.076 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=129 \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=128 \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=128 \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:42:42 +0000 (0:00:02.803)       0:11:08.880 ******** \n=============================================================================== \nloadbalancer : Start backup proxysql container ------------------------- 13.03s\nloadbalancer : Start master proxysql container ------------------------- 12.64s\nloadbalancer : Start backup keepalived container ------------------------ 7.75s\nhaproxy-config : Copying over skyline haproxy config -------------------- 7.72s\nloadbalancer : Start backup haproxy container --------------------------- 7.61s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 7.54s\nloadbalancer : Stop backup haproxy container ---------------------------- 7.44s\nhaproxy-config : Copying over nova haproxy config ----------------------- 7.27s\nloadbalancer : Stop backup proxysql container --------------------------- 7.14s\nloadbalancer : Start master haproxy container --------------------------- 7.03s\nloadbalancer : Stop backup keepalived container ------------------------- 7.03s\nloadbalancer : Stop master keepalived container ------------------------- 6.96s\nloadbalancer : Stop master haproxy container ---------------------------- 6.84s\nloadbalancer : Stop master proxysql container --------------------------- 6.80s\nloadbalancer : Start master keepalived container ------------------------ 6.74s\nhaproxy-config : Copying over neutron haproxy config -------------------- 6.37s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 6.03s\nhaproxy-config : Copying over glance haproxy config --------------------- 5.71s\nloadbalancer : Copying checks for services which are enabled ------------ 5.08s\nhaproxy-config : Copying over manila haproxy config --------------------- 4.92s\n+ osism apply -a upgrade opensearch\n\u001b[32m2026-03-26 04:42:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 880f3991-9db2-4938-9bc9-5c310c65b633 (opensearch) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:42:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 880f3991-9db2-4938-9bc9-5c310c65b633 (opensearch) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:42:51 +0000 (0:00:01.527)       0:00:01.527 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:42:53 +0000 (0:00:01.660)       0:00:03.188 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  04:42:55 +0000 (0:00:02.422)       0:00:05.611 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nThursday 26 March 2026  04:42:57 +0000 (0:00:02.344)       0:00:07.955 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nThursday 26 March 2026  04:42:59 +0000 (0:00:02.096)       0:00:10.052 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  04:43:02 +0000 (0:00:02.431)       0:00:12.483 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nThursday 26 March 2026  04:43:04 +0000 (0:00:01.804)       0:00:14.287 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:43:07 +0000 (0:00:03.512)       0:00:17.800 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nThursday 26 March 2026  04:43:09 +0000 (0:00:01.824)       0:00:19.625 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nThursday 26 March 2026  04:43:11 +0000 (0:00:01.823)       0:00:21.449 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nThursday 26 March 2026  04:43:15 +0000 (0:00:03.883)       0:00:25.333 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nThursday 26 March 2026  04:43:18 +0000 (0:00:03.433)       0:00:28.766 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : opensearch | Check containers] ****************\nThursday 26 March 2026  04:43:21 +0000 (0:00:03.101)       0:00:31.867 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [service-check-containers : opensearch | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:43:25 +0000 (0:00:03.627)       0:00:35.495 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:43:26 +0000 (0:00:01.361)       0:00:36.856 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  04:43:28 +0000 (0:00:02.107)       0:00:38.964 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  04:43:30 +0000 (0:00:01.662)       0:00:40.627 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  04:43:30 +0000 (0:00:00.475)       0:00:41.103 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nThursday 26 March 2026  04:43:31 +0000 (0:00:00.445)       0:00:41.549 ******** \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nThursday 26 March 2026  04:43:32 +0000 (0:00:00.853)       0:00:42.402 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nThursday 26 March 2026  04:43:35 +0000 (0:00:03.592)       0:00:45.994 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nThursday 26 March 2026  04:43:40 +0000 (0:00:04.453)       0:00:50.448 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nThursday 26 March 2026  04:44:49 +0000 (0:01:09.534)       0:01:59.982 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nThursday 26 March 2026  04:46:20 +0000 (0:01:30.495)       0:03:30.478 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nThursday 26 March 2026  04:46:22 +0000 (0:00:01.688)       0:03:32.166 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nThursday 26 March 2026  04:46:25 +0000 (0:00:03.316)       0:03:35.483 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nThursday 26 March 2026  04:46:28 +0000 (0:00:03.285)       0:03:38.768 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nThursday 26 March 2026  04:46:29 +0000 (0:00:01.249)       0:03:40.017 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:46:31 +0000 (0:00:01.996)       0:03:42.014 ******** \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 90.50s\nopensearch : Restart opensearch container ------------------------------ 69.53s\nopensearch : Perform a flush -------------------------------------------- 4.45s\nopensearch : Copying over config.json files for services ---------------- 3.88s\nservice-check-containers : opensearch | Check containers ---------------- 3.63s\nopensearch : Disable shard allocation ----------------------------------- 3.59s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 3.51s\nopensearch : Copying over opensearch service config file ---------------- 3.43s\nopensearch : Wait for OpenSearch to become ready ------------------------ 3.32s\nopensearch : Check if a log retention policy exists --------------------- 3.29s\nopensearch : Copying over opensearch-dashboards config file ------------- 3.10s\nopensearch : Ensuring config directories exist -------------------------- 2.43s\nGroup hosts based on enabled services ----------------------------------- 2.42s\nopensearch : include_tasks ---------------------------------------------- 2.34s\nservice-check-containers : Include tasks -------------------------------- 2.11s\nopensearch : Setting sysctl values -------------------------------------- 2.10s\nopensearch : Apply retention policy to existing indices ----------------- 2.00s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.82s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 1.82s\nopensearch : include_tasks ---------------------------------------------- 1.80s\n+ osism apply -a upgrade memcached\n\u001b[32m2026-03-26 04:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08a40978-d999-48b6-a67e-60a08e1bdf47 (memcached) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08a40978-d999-48b6-a67e-60a08e1bdf47 (memcached) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x758c9b818c20>): Expecting value: line 2 column 1 (char 1)\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x758c9b818c20>): 'NoneType' object is not subscriptable\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:46:39 +0000 (0:00:01.066)       0:00:01.066 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:46:40 +0000 (0:00:00.952)       0:00:02.018 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nThursday 26 March 2026  04:46:41 +0000 (0:00:00.878)       0:00:02.897 ******** \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nThursday 26 March 2026  04:46:43 +0000 (0:00:01.599)       0:00:04.496 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nThursday 26 March 2026  04:46:44 +0000 (0:00:00.931)       0:00:05.428 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [service-check-containers : memcached | Check containers] *****************\nThursday 26 March 2026  04:46:46 +0000 (0:00:01.886)       0:00:07.315 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [service-check-containers : memcached | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:46:47 +0000 (0:00:01.247)       0:00:08.562 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:46:47 +0000 (0:00:00.353)       0:00:08.916 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m\n\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x758c9b818c20>): 'NoneType' object is not subscriptable\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nThursday 26 March 2026  04:46:49 +0000 (0:00:01.342)       0:00:10.259 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:46:59 +0000 (0:00:10.074)       0:00:20.334 ******** \n=============================================================================== \nmemcached : Restart memcached container -------------------------------- 10.08s\nmemcached : Copying over config.json files for services ----------------- 1.89s\nmemcached : include_tasks ----------------------------------------------- 1.60s\nservice-check-containers : Include tasks -------------------------------- 1.34s\nservice-check-containers : memcached | Check containers ----------------- 1.25s\nGroup hosts based on Kolla action --------------------------------------- 0.95s\nmemcached : Ensuring config directories exist --------------------------- 0.93s\nGroup hosts based on enabled services ----------------------------------- 0.88s\nservice-check-containers : memcached | Notify handlers to restart containers --- 0.35s\n+ osism apply -a upgrade redis\n\u001b[32m2026-03-26 04:47:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 103a9efa-0088-4a4c-9ba5-5bed8cc516cc (redis) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:47:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 103a9efa-0088-4a4c-9ba5-5bed8cc516cc (redis) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:47:07 +0000 (0:00:01.360)       0:00:01.360 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:47:09 +0000 (0:00:02.088)       0:00:03.449 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nThursday 26 March 2026  04:47:12 +0000 (0:00:02.799)       0:00:06.248 ******** \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nThursday 26 March 2026  04:47:14 +0000 (0:00:01.728)       0:00:07.977 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nThursday 26 March 2026  04:47:16 +0000 (0:00:02.252)       0:00:10.229 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nThursday 26 March 2026  04:47:19 +0000 (0:00:03.169)       0:00:13.399 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : redis | Check containers] *********************\nThursday 26 March 2026  04:47:23 +0000 (0:00:03.891)       0:00:17.291 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : redis | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:47:26 +0000 (0:00:03.135)       0:00:20.426 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:47:28 +0000 (0:00:01.612)       0:00:22.038 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  04:47:30 +0000 (0:00:01.925)       0:00:23.964 ******** \n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  04:47:30 +0000 (0:00:00.466)       0:00:24.430 ******** \n\nTASK [redis : Flush handlers] **************************************************\nThursday 26 March 2026  04:47:31 +0000 (0:00:00.450)       0:00:24.881 ******** \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nThursday 26 March 2026  04:47:32 +0000 (0:00:00.802)       0:00:25.684 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nThursday 26 March 2026  04:47:43 +0000 (0:00:10.979)       0:00:36.663 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:47:53 +0000 (0:00:10.908)       0:00:47.572 ******** \n=============================================================================== \nredis : Restart redis container ---------------------------------------- 10.98s\nredis : Restart redis-sentinel container ------------------------------- 10.91s\nredis : Copying over redis config files --------------------------------- 3.89s\nredis : Copying over default config.json files -------------------------- 3.17s\nservice-check-containers : redis | Check containers --------------------- 3.14s\nGroup hosts based on enabled services ----------------------------------- 2.80s\nredis : Ensuring config directories exist ------------------------------- 2.25s\nGroup hosts based on Kolla action --------------------------------------- 2.09s\nservice-check-containers : Include tasks -------------------------------- 1.93s\nredis : include_tasks --------------------------------------------------- 1.73s\nredis : Flush handlers -------------------------------------------------- 1.72s\nservice-check-containers : redis | Notify handlers to restart containers --- 1.61s\n+ osism apply -a upgrade mariadb\n\u001b[32m2026-03-26 04:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44634b2e-5b51-4f9a-9bb1-cb1a991037c4 (mariadb) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 44634b2e-5b51-4f9a-9bb1-cb1a991037c4 (mariadb) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:48:02 +0000 (0:00:01.330)       0:00:01.330 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:48:04 +0000 (0:00:02.378)       0:00:03.708 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nThursday 26 March 2026  04:48:07 +0000 (0:00:02.678)       0:00:06.387 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  04:48:08 +0000 (0:00:01.508)       0:00:07.895 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nThursday 26 March 2026  04:48:10 +0000 (0:00:01.753)       0:00:09.649 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nThursday 26 March 2026  04:48:14 +0000 (0:00:03.606)       0:00:13.255 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nThursday 26 March 2026  04:48:15 +0000 (0:00:01.606)       0:00:14.862 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nThursday 26 March 2026  04:48:18 +0000 (0:00:02.251)       0:00:17.116 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nThursday 26 March 2026  04:48:22 +0000 (0:00:04.504)       0:00:21.621 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nThursday 26 March 2026  04:48:24 +0000 (0:00:02.049)       0:00:23.671 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  04:48:29 +0000 (0:00:04.773)       0:00:28.444 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nThursday 26 March 2026  04:48:31 +0000 (0:00:01.881)       0:00:30.326 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:48:34 +0000 (0:00:03.572)       0:00:33.899 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nThursday 26 March 2026  04:48:38 +0000 (0:00:03.520)       0:00:37.420 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : mariadb | Check containers] *******************\nThursday 26 March 2026  04:48:42 +0000 (0:00:04.149)       0:00:41.569 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [service-check-containers : mariadb | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:48:46 +0000 (0:00:04.154)       0:00:45.723 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:48:48 +0000 (0:00:01.392)       0:00:47.116 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Checking for mariadb cluster] **********************************\nThursday 26 March 2026  04:48:52 +0000 (0:00:04.186)       0:00:51.302 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Cleaning up temp file on localhost] ****************************\nThursday 26 March 2026  04:48:53 +0000 (0:00:01.400)       0:00:52.703 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Stop MariaDB containers] ***************************************\nThursday 26 March 2026  04:48:54 +0000 (0:00:01.131)       0:00:53.835 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Run MariaDB wsrep recovery] ************************************\nThursday 26 March 2026  04:48:56 +0000 (0:00:01.391)       0:00:55.226 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Copying MariaDB log file to /tmp] ******************************\nThursday 26 March 2026  04:48:57 +0000 (0:00:01.536)       0:00:56.763 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Get MariaDB wsrep recovery seqno] ******************************\nThursday 26 March 2026  04:48:59 +0000 (0:00:01.358)       0:00:58.122 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Removing MariaDB log file from /tmp] ***************************\nThursday 26 March 2026  04:49:01 +0000 (0:00:01.927)       0:01:00.050 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Registering MariaDB seqno variable] ****************************\nThursday 26 March 2026  04:49:02 +0000 (0:00:01.357)       0:01:01.407 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Comparing seqno value on all mariadb hosts] ********************\nThursday 26 March 2026  04:49:03 +0000 (0:00:01.545)       0:01:02.953 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Writing hostname of host with the largest seqno to temp file] ***\nThursday 26 March 2026  04:49:05 +0000 (0:00:01.507)       0:01:04.460 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Registering mariadb_recover_inventory_name from temp file] *****\nThursday 26 March 2026  04:49:06 +0000 (0:00:01.368)       0:01:05.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Store bootstrap and master hostnames into facts] ***************\nThursday 26 March 2026  04:49:08 +0000 (0:00:01.405)       0:01:07.234 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Set grastate.dat file from MariaDB container in bootstrap host] ***\nThursday 26 March 2026  04:49:09 +0000 (0:00:01.336)       0:01:08.571 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Starting first MariaDB container] ******************************\nThursday 26 March 2026  04:49:10 +0000 (0:00:01.377)       0:01:09.949 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for first MariaDB container] ******************************\nThursday 26 March 2026  04:49:12 +0000 (0:00:01.332)       0:01:11.281 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Set first MariaDB container as primary] ************************\nThursday 26 March 2026  04:49:13 +0000 (0:00:01.568)       0:01:12.850 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB to become operational] ************************\nThursday 26 March 2026  04:49:15 +0000 (0:00:01.475)       0:01:14.325 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Restart slave MariaDB container(s)] ****************************\nThursday 26 March 2026  04:49:16 +0000 (0:00:01.400)       0:01:15.726 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for slave MariaDB] ****************************************\nThursday 26 March 2026  04:49:20 +0000 (0:00:03.569)       0:01:19.295 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Restart master MariaDB container(s)] ***************************\nThursday 26 March 2026  04:49:21 +0000 (0:00:01.582)       0:01:20.878 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for master mariadb] ***************************************\nThursday 26 March 2026  04:49:25 +0000 (0:00:03.409)       0:01:24.287 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Get container facts] ***************************\nThursday 26 March 2026  04:49:26 +0000 (0:00:01.730)       0:01:26.018 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are missing or not running] ***\nThursday 26 March 2026  04:49:28 +0000 (0:00:01.362)       0:01:27.380 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are unhealthy] **************\nThursday 26 March 2026  04:49:29 +0000 (0:00:01.330)       0:01:28.711 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nThursday 26 March 2026  04:49:31 +0000 (0:00:01.801)       0:01:30.512 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nThursday 26 March 2026  04:49:33 +0000 (0:00:01.951)       0:01:32.464 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nThursday 26 March 2026  04:49:35 +0000 (0:00:01.959)       0:01:34.423 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nThursday 26 March 2026  04:49:36 +0000 (0:00:01.336)       0:01:35.760 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nThursday 26 March 2026  04:49:38 +0000 (0:00:01.387)       0:01:37.147 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nThursday 26 March 2026  04:49:40 +0000 (0:00:02.028)       0:01:39.176 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nThursday 26 March 2026  04:49:41 +0000 (0:00:01.443)       0:01:40.619 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nThursday 26 March 2026  04:49:42 +0000 (0:00:01.379)       0:01:41.999 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nThursday 26 March 2026  04:49:46 +0000 (0:00:03.524)       0:01:45.523 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nThursday 26 March 2026  04:49:47 +0000 (0:00:01.398)       0:01:46.921 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nThursday 26 March 2026  04:49:49 +0000 (0:00:01.455)       0:01:48.377 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  04:49:51 +0000 (0:00:01.793)       0:01:50.171 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nThursday 26 March 2026  04:49:52 +0000 (0:00:01.537)       0:01:51.709 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nThursday 26 March 2026  04:49:54 +0000 (0:00:01.619)       0:01:53.329 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nThursday 26 March 2026  04:49:55 +0000 (0:00:01.685)       0:01:55.014 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  04:49:58 +0000 (0:00:02.231)       0:01:57.246 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  04:50:24 +0000 (0:00:26.258)       0:02:23.505 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  04:50:29 +0000 (0:00:04.617)       0:02:28.123 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  04:50:32 +0000 (0:00:02.976)       0:02:31.099 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  04:50:58 +0000 (0:00:26.022)       0:02:57.122 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  04:51:02 +0000 (0:00:04.564)       0:03:01.686 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nThursday 26 March 2026  04:51:05 +0000 (0:00:02.870)       0:03:04.557 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nThursday 26 March 2026  04:51:30 +0000 (0:00:25.220)       0:03:29.777 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Wait for MariaDB service port liveness (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nThursday 26 March 2026  04:51:38 +0000 (0:00:07.967)       0:03:37.745 ******** \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nThursday 26 March 2026  04:51:42 +0000 (0:00:03.948)       0:03:41.693 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nThursday 26 March 2026  04:51:44 +0000 (0:00:01.988)       0:03:43.681 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nThursday 26 March 2026  04:51:47 +0000 (0:00:03.098)       0:03:46.780 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nThursday 26 March 2026  04:51:51 +0000 (0:00:03.287)       0:03:50.067 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nThursday 26 March 2026  04:51:54 +0000 (0:00:03.181)       0:03:53.248 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-check : mariadb | Get container facts] ***************************\nThursday 26 March 2026  04:51:57 +0000 (0:00:03.482)       0:03:56.731 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are missing or not running] ***\nThursday 26 March 2026  04:52:04 +0000 (0:00:06.649)       0:04:03.381 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are unhealthy] **************\nThursday 26 March 2026  04:52:08 +0000 (0:00:03.812)       0:04:07.193 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nThursday 26 March 2026  04:52:09 +0000 (0:00:01.604)       0:04:08.798 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nThursday 26 March 2026  04:52:13 +0000 (0:00:03.415)       0:04:12.214 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Run upgrade in MariaDB container] ******************************\nThursday 26 March 2026  04:52:15 +0000 (0:00:01.996)       0:04:14.210 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:52:32 +0000 (0:00:17.715)       0:04:31.925 ******** \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 77.50s\nmariadb : Run upgrade in MariaDB container ----------------------------- 17.72s\nmariadb : Wait for MariaDB service port liveness ----------------------- 17.15s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 9.79s\nservice-check : mariadb | Get container facts --------------------------- 6.65s\nmariadb : Copying over galera.cnf --------------------------------------- 4.77s\nmariadb : Copying over config.json files for services ------------------- 4.51s\nservice-check-containers : Include tasks -------------------------------- 4.19s\nservice-check-containers : mariadb | Check containers ------------------- 4.15s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 4.15s\nservice-check : mariadb | Fail if containers are missing or not running --- 3.81s\nmariadb : Ensuring config directories exist ----------------------------- 3.61s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 3.57s\nmariadb : Restart slave MariaDB container(s) ---------------------------- 3.57s\nmariadb : Check MariaDB service WSREP sync status ----------------------- 3.52s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 3.52s\nmariadb : Granting permissions on Mariabackup database to backup user --- 3.48s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 3.42s\nmariadb : Restart master MariaDB container(s) --------------------------- 3.41s\nmariadb : Creating mysql monitor user ----------------------------------- 3.29s\n+ osism apply -a upgrade rabbitmq\n\u001b[32m2026-03-26 04:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e3e4e19-e6c3-4dbe-b118-c2b62a5d823d (rabbitmq) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2e3e4e19-e6c3-4dbe-b118-c2b62a5d823d (rabbitmq) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:52:41 +0000 (0:00:01.444)       0:00:01.444 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:52:43 +0000 (0:00:01.871)       0:00:03.315 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  04:52:45 +0000 (0:00:01.751)       0:00:05.067 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nThursday 26 March 2026  04:52:47 +0000 (0:00:02.804)       0:00:07.871 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nThursday 26 March 2026  04:52:50 +0000 (0:00:02.382)       0:00:10.254 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nThursday 26 March 2026  04:52:53 +0000 (0:00:03.398)       0:00:13.653 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nThursday 26 March 2026  04:53:02 +0000 (0:00:09.086)       0:00:22.740 ******** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nThursday 26 March 2026  04:53:04 +0000 (0:00:01.301)       0:00:24.041 ******** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  04:53:05 +0000 (0:00:01.725)       0:00:25.767 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nThursday 26 March 2026  04:53:07 +0000 (0:00:01.906)       0:00:27.674 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nThursday 26 March 2026  04:53:09 +0000 (0:00:02.306)       0:00:29.981 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nThursday 26 March 2026  04:53:12 +0000 (0:00:02.750)       0:00:32.731 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nThursday 26 March 2026  04:53:14 +0000 (0:00:01.877)       0:00:34.609 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nThursday 26 March 2026  04:53:16 +0000 (0:00:01.771)       0:00:36.380 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nThursday 26 March 2026  04:53:18 +0000 (0:00:02.399)       0:00:38.780 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nThursday 26 March 2026  04:53:21 +0000 (0:00:02.454)       0:00:41.235 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nThursday 26 March 2026  04:53:24 +0000 (0:00:03.109)       0:00:44.344 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nThursday 26 March 2026  04:53:26 +0000 (0:00:02.426)       0:00:46.770 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nThursday 26 March 2026  04:53:29 +0000 (0:00:02.314)       0:00:49.085 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nThursday 26 March 2026  04:53:31 +0000 (0:00:02.290)       0:00:51.376 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nThursday 26 March 2026  04:53:33 +0000 (0:00:02.591)       0:00:53.968 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over extra CA certificates] *******\nThursday 26 March 2026  04:53:35 +0000 (0:00:01.793)       0:00:55.761 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over backend internal TLS certificate] ***\nThursday 26 March 2026  04:53:38 +0000 (0:00:02.341)       0:00:58.103 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over backend internal TLS key] ****\nThursday 26 March 2026  04:53:39 +0000 (0:00:01.452)       0:00:59.555 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nThursday 26 March 2026  04:53:41 +0000 (0:00:01.778)       0:01:01.334 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : rabbitmq | Check containers] ******************\nThursday 26 March 2026  04:53:45 +0000 (0:00:03.710)       0:01:05.044 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [service-check-containers : rabbitmq | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:53:47 +0000 (0:00:02.187)       0:01:07.231 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:53:48 +0000 (0:00:01.416)       0:01:08.648 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nThursday 26 March 2026  04:53:50 +0000 (0:00:02.091)       0:01:10.740 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  04:53:52 +0000 (0:00:01.778)       0:01:12.518 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  04:53:54 +0000 (0:00:02.163)       0:01:14.682 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  04:54:03 +0000 (0:00:09.146)       0:01:23.828 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  04:54:12 +0000 (0:00:09.086)       0:01:32.915 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  04:54:21 +0000 (0:00:08.807)       0:01:41.722 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  04:54:23 +0000 (0:00:01.683)       0:01:43.405 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  04:54:31 +0000 (0:00:08.028)       0:01:51.434 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  04:54:45 +0000 (0:00:13.625)       0:02:05.060 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nThursday 26 March 2026  04:54:53 +0000 (0:00:08.243)       0:02:13.303 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nThursday 26 March 2026  04:54:55 +0000 (0:00:01.729)       0:02:15.033 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nThursday 26 March 2026  04:55:03 +0000 (0:00:08.941)       0:02:23.974 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nThursday 26 March 2026  04:55:18 +0000 (0:00:14.227)       0:02:38.202 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nThursday 26 March 2026  04:55:28 +0000 (0:00:10.661)       0:02:48.864 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nThursday 26 March 2026  04:55:30 +0000 (0:00:01.517)       0:02:50.382 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:55:34 +0000 (0:00:04.463)       0:02:54.846 ******** \n=============================================================================== \nrabbitmq : Restart rabbitmq container ---------------------------------- 36.94s\nrabbitmq : Waiting for rabbitmq to start ------------------------------- 27.71s\nrabbitmq : Put RabbitMQ node into maintenance mode --------------------- 26.12s\nrabbitmq : Get new RabbitMQ version ------------------------------------- 9.09s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 5.58s\nrabbitmq : Enable all stable feature flags ------------------------------ 4.46s\nrabbitmq : Enable all stable feature flags ------------------------------ 3.71s\nrabbitmq : Get current RabbitMQ version --------------------------------- 3.40s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 3.11s\nrabbitmq : include_tasks ------------------------------------------------ 2.80s\nrabbitmq : List RabbitMQ policies --------------------------------------- 2.75s\nrabbitmq : Copying over enabled_plugins --------------------------------- 2.59s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 2.45s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 2.43s\nrabbitmq : Copying over config.json files for services ------------------ 2.40s\nrabbitmq : Get container facts ------------------------------------------ 2.38s\nservice-cert-copy : rabbitmq | Copying over extra CA certificates ------- 2.34s\nrabbitmq : Copying over advanced.config --------------------------------- 2.32s\nrabbitmq : Get container facts ------------------------------------------ 2.31s\nrabbitmq : Copying over definitions.json -------------------------------- 2.29s\n+ osism apply -a upgrade openvswitch\n\u001b[32m2026-03-26 04:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3fbc17bd-68e1-4e2f-bde2-7bb54783be9e (openvswitch) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3fbc17bd-68e1-4e2f-bde2-7bb54783be9e (openvswitch) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x739f147b4c20>): Expecting value: line 2 column 1 (char 1)\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x739f147b4c20>): 'NoneType' object is not subscriptable\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:55:42 +0000 (0:00:01.079)       0:00:01.079 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:55:44 +0000 (0:00:01.553)       0:00:02.633 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nThursday 26 March 2026  04:55:45 +0000 (0:00:01.072)       0:00:03.706 ******** \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nThursday 26 March 2026  04:55:47 +0000 (0:00:02.150)       0:00:05.856 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nThursday 26 March 2026  04:55:49 +0000 (0:00:01.451)       0:00:07.308 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nThursday 26 March 2026  04:55:50 +0000 (0:00:01.386)       0:00:08.694 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nThursday 26 March 2026  04:55:52 +0000 (0:00:01.959)       0:00:10.654 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nThursday 26 March 2026  04:55:53 +0000 (0:00:01.091)       0:00:11.746 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nThursday 26 March 2026  04:55:55 +0000 (0:00:01.717)       0:00:13.463 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nThursday 26 March 2026  04:55:57 +0000 (0:00:02.390)       0:00:15.854 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-check-containers : openvswitch | Check containers] ***************\nThursday 26 March 2026  04:55:59 +0000 (0:00:01.442)       0:00:17.296 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : openvswitch | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:56:01 +0000 (0:00:02.097)       0:00:19.393 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:56:02 +0000 (0:00:00.936)       0:00:20.330 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m\n\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x739f147b4c20>): 'NoneType' object is not subscriptable\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:03 +0000 (0:00:01.833)       0:00:22.164 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:04 +0000 (0:00:00.164)       0:00:22.329 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:04 +0000 (0:00:00.153)       0:00:22.483 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:04 +0000 (0:00:00.144)       0:00:22.627 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:04 +0000 (0:00:00.348)       0:00:22.975 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nThursday 26 March 2026  04:56:04 +0000 (0:00:00.145)       0:00:23.121 ******** \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nThursday 26 March 2026  04:56:05 +0000 (0:00:00.150)       0:00:23.271 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nThursday 26 March 2026  04:56:15 +0000 (0:00:10.909)       0:00:34.180 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nThursday 26 March 2026  04:56:17 +0000 (0:00:01.206)       0:00:35.387 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nThursday 26 March 2026  04:56:27 +0000 (0:00:10.486)       0:00:45.873 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nThursday 26 March 2026  04:56:34 +0000 (0:00:06.323)       0:00:52.197 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=br-ex)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nThursday 26 March 2026  04:56:36 +0000 (0:00:02.279)       0:00:54.476 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  04:56:40 +0000 (0:00:04.022)       0:00:58.498 ******** \n=============================================================================== \nopenvswitch : Restart openvswitch-db-server container ------------------ 10.91s\nopenvswitch : Restart openvswitch-vswitchd container ------------------- 10.49s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 6.32s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 4.02s\nopenvswitch : Copying over config.json files for services --------------- 2.39s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 2.28s\nopenvswitch : include_tasks --------------------------------------------- 2.15s\nservice-check-containers : openvswitch | Check containers --------------- 2.10s\nmodule-load : Drop module persistence ----------------------------------- 1.96s\nservice-check-containers : Include tasks -------------------------------- 1.83s\nopenvswitch : Ensuring config directories exist ------------------------- 1.72s\nGroup hosts based on Kolla action --------------------------------------- 1.55s\nmodule-load : Load modules ---------------------------------------------- 1.45s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 1.44s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.39s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 1.21s\nopenvswitch : Flush Handlers -------------------------------------------- 1.11s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 1.09s\nGroup hosts based on enabled services ----------------------------------- 1.07s\nservice-check-containers : openvswitch | Notify handlers to restart containers --- 0.94s\n+ osism apply -a upgrade ovn\n\u001b[32m2026-03-26 04:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 53508c48-b0a1-45e6-8522-6a5123067a74 (ovn) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 04:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 53508c48-b0a1-45e6-8522-6a5123067a74 (ovn) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nThursday 26 March 2026  04:56:48 +0000 (0:00:01.326)       0:00:01.326 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nThursday 26 March 2026  04:56:52 +0000 (0:00:03.418)       0:00:04.744 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nThursday 26 March 2026  04:56:55 +0000 (0:00:03.073)       0:00:07.818 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nThursday 26 March 2026  04:56:58 +0000 (0:00:02.771)       0:00:10.590 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nThursday 26 March 2026  04:57:00 +0000 (0:00:02.402)       0:00:12.993 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nThursday 26 March 2026  04:57:03 +0000 (0:00:02.655)       0:00:15.648 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nThursday 26 March 2026  04:57:05 +0000 (0:00:02.288)       0:00:17.937 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_controller | Check containers] ************\nThursday 26 March 2026  04:57:08 +0000 (0:00:03.144)       0:00:21.081 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_controller | Notify handlers to restart containers] ***\nThursday 26 March 2026  04:57:11 +0000 (0:00:02.529)       0:00:23.611 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  04:57:13 +0000 (0:00:02.009)       0:00:25.621 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nThursday 26 March 2026  04:57:15 +0000 (0:00:02.506)       0:00:28.127 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nThursday 26 March 2026  04:57:19 +0000 (0:00:03.550)       0:00:31.677 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:39 +0000 (0:00:20.184)       0:00:51.861 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:39 +0000 (0:00:00.458)       0:00:52.320 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:40 +0000 (0:00:00.464)       0:00:52.785 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:40 +0000 (0:00:00.445)       0:00:53.231 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:41 +0000 (0:00:00.450)       0:00:53.681 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nThursday 26 March 2026  04:57:41 +0000 (0:00:00.457)       0:00:54.139 ******** \n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nThursday 26 March 2026  04:57:42 +0000 (0:00:00.841)       0:00:54.980 ******** \n\nSTILL ALIVE [task 'ovn-controller : Restart ovn-controller container' is running] ***\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  04:59:54 +0000 (0:02:11.773)       0:03:06.754 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  04:59:56 +0000 (0:00:01.950)       0:03:08.705 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nThursday 26 March 2026  04:59:58 +0000 (0:00:01.933)       0:03:10.639 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nThursday 26 March 2026  04:59:59 +0000 (0:00:01.802)       0:03:12.442 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nThursday 26 March 2026  05:00:01 +0000 (0:00:01.355)       0:03:13.797 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nThursday 26 March 2026  05:00:02 +0000 (0:00:01.418)       0:03:15.216 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nThursday 26 March 2026  05:00:04 +0000 (0:00:01.650)       0:03:16.866 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nThursday 26 March 2026  05:00:05 +0000 (0:00:01.423)       0:03:18.290 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nThursday 26 March 2026  05:00:07 +0000 (0:00:01.482)       0:03:19.772 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nThursday 26 March 2026  05:00:08 +0000 (0:00:01.687)       0:03:21.459 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nThursday 26 March 2026  05:00:10 +0000 (0:00:01.505)       0:03:22.965 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nThursday 26 March 2026  05:00:12 +0000 (0:00:01.863)       0:03:24.828 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nThursday 26 March 2026  05:00:13 +0000 (0:00:01.432)       0:03:26.260 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nThursday 26 March 2026  05:00:15 +0000 (0:00:01.333)       0:03:27.594 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nThursday 26 March 2026  05:00:16 +0000 (0:00:01.425)       0:03:29.019 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nThursday 26 March 2026  05:00:18 +0000 (0:00:01.814)       0:03:30.834 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nThursday 26 March 2026  05:00:19 +0000 (0:00:01.345)       0:03:32.180 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nThursday 26 March 2026  05:00:21 +0000 (0:00:02.167)       0:03:34.348 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nThursday 26 March 2026  05:00:23 +0000 (0:00:01.441)       0:03:35.790 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nThursday 26 March 2026  05:00:24 +0000 (0:00:01.411)       0:03:37.201 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nThursday 26 March 2026  05:00:26 +0000 (0:00:01.722)       0:03:38.923 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nThursday 26 March 2026  05:00:30 +0000 (0:00:04.227)       0:03:43.151 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Ensure configuration for relays exists] *************************\nThursday 26 March 2026  05:00:37 +0000 (0:00:06.531)       0:03:49.683 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****\nThursday 26 March 2026  05:00:39 +0000 (0:00:01.944)       0:03:51.627 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for OVN relay services] **********\nThursday 26 March 2026  05:00:40 +0000 (0:00:01.864)       0:03:53.492 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Generate config files for OVN relay services] *******************\nThursday 26 March 2026  05:00:43 +0000 (0:00:02.670)       0:03:56.163 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers] ********************\nThursday 26 March 2026  05:00:46 +0000 (0:00:03.022)       0:03:59.186 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nThursday 26 March 2026  05:00:52 +0000 (0:00:05.498)       0:04:04.684 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nThursday 26 March 2026  05:00:53 +0000 (0:00:01.527)       0:04:06.212 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers with iteration] *****\nThursday 26 March 2026  05:00:57 +0000 (0:00:03.304)       0:04:09.517 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nThursday 26 March 2026  05:00:59 +0000 (0:00:02.246)       0:04:11.764 ******** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  05:01:00 +0000 (0:00:01.561)       0:04:13.325 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  05:01:01 +0000 (0:00:00.424)       0:04:13.750 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nThursday 26 March 2026  05:01:01 +0000 (0:00:00.453)       0:04:14.203 ******** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nThursday 26 March 2026  05:01:02 +0000 (0:00:01.021)       0:04:15.225 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nThursday 26 March 2026  05:01:19 +0000 (0:00:16.856)       0:04:32.081 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************\nThursday 26 March 2026  05:01:36 +0000 (0:00:16.931)       0:04:49.013 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nThursday 26 March 2026  05:01:47 +0000 (0:00:11.413)       0:05:00.426 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nThursday 26 March 2026  05:02:05 +0000 (0:00:17.732)       0:05:18.159 ******** \nPausing for 5 seconds\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nThursday 26 March 2026  05:02:11 +0000 (0:00:06.175)       0:05:24.334 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nThursday 26 March 2026  05:02:13 +0000 (0:00:01.883)       0:05:26.217 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nThursday 26 March 2026  05:02:15 +0000 (0:00:01.825)       0:05:28.043 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nThursday 26 March 2026  05:02:17 +0000 (0:00:01.865)       0:05:29.908 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nThursday 26 March 2026  05:02:19 +0000 (0:00:01.640)       0:05:31.549 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nThursday 26 March 2026  05:02:20 +0000 (0:00:01.938)       0:05:33.487 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************\nThursday 26 March 2026  05:02:22 +0000 (0:00:01.912)       0:05:35.399 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=49  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=48  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=47  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  05:02:26 +0000 (0:00:03.159)       0:05:38.559 ******** \n=============================================================================== \novn-controller : Restart ovn-controller container --------------------- 131.77s\novn-controller : Configure OVN in OVSDB -------------------------------- 20.18s\novn-db : Restart ovn-northd container ---------------------------------- 17.73s\novn-db : Restart ovn-sb-db container ----------------------------------- 16.93s\novn-db : Restart ovn-nb-db container ----------------------------------- 16.86s\novn-db : Restart ovn-sb-db-relay container ----------------------------- 11.41s\novn-db : Copying over config.json files for services -------------------- 6.53s\novn-db : Wait for leader election --------------------------------------- 6.18s\nservice-check-containers : ovn_db | Check containers -------------------- 5.50s\novn-db : Ensuring config directories exist ------------------------------ 4.23s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 3.55s\nGroup hosts based on Kolla action --------------------------------------- 3.42s\nservice-check-containers : Include tasks -------------------------------- 3.30s\novn-db : Wait for ovn-sb-db-relay --------------------------------------- 3.16s\novn-controller : Copying over systemd override -------------------------- 3.14s\novn-controller : Flush handlers ----------------------------------------- 3.12s\nGroup hosts based on enabled services ----------------------------------- 3.07s\novn-db : Generate config files for OVN relay services ------------------- 3.02s\novn-controller : include_tasks ------------------------------------------ 2.77s\novn-db : Copying over config.json files for OVN relay services ---------- 2.67s\n+ [[ false == \\f\\a\\l\\s\\e ]]\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ osism apply ceph-rolling_update -e ireallymeanit=yes\n\u001b[32m2026-03-26 05:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3cc48515-9882-4eae-9b27-8ff8065ea8e3 (ceph-rolling_update) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 05:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3cc48515-9882-4eae-9b27-8ff8065ea8e3 (ceph-rolling_update) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Confirm whether user really meant to upgrade the cluster] ****************\n\nTASK [Exit playbook, if user did not mean to upgrade cluster] ******************\nThursday 26 March 2026  05:02:38 +0000 (0:00:01.792)       0:00:01.792 ******** \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: rbdmirrors\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: nfss\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: clients\u001b[0m\n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nPLAY [Gather facts and check the init system] **********************************\n\nTASK [Gather facts on all Ceph hosts for following reference] ******************\nThursday 26 March 2026  05:02:39 +0000 (0:00:01.712)       0:00:03.505 ******** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather facts] ************************************************************\nThursday 26 March 2026  05:02:47 +0000 (0:00:07.308)       0:00:10.813 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Gather and delegate facts] ***********************************************\nThursday 26 March 2026  05:02:53 +0000 (0:00:06.298)       0:00:17.111 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [Set_fact rolling_update] *************************************************\nThursday 26 March 2026  05:03:24 +0000 (0:00:31.499)       0:00:48.611 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:03:27 +0000 (0:00:02.108)       0:00:50.719 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:03:29 +0000 (0:00:02.740)       0:00:53.460 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:03:32 +0000 (0:00:02.663)       0:00:56.123 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:03:34 +0000 (0:00:01.965)       0:00:58.089 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:03:36 +0000 (0:00:02.538)       0:01:00.627 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:03:38 +0000 (0:00:01.987)       0:01:02.615 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:03:41 +0000 (0:00:02.137)       0:01:04.752 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:03:43 +0000 (0:00:01.905)       0:01:06.658 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:03:45 +0000 (0:00:02.534)       0:01:09.192 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:03:47 +0000 (0:00:02.237)       0:01:11.430 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:03:49 +0000 (0:00:01.661)       0:01:13.092 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:03:51 +0000 (0:00:02.119)       0:01:15.211 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:03:54 +0000 (0:00:03.307)       0:01:18.519 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:03:56 +0000 (0:00:01.437)       0:01:19.956 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:03:58 +0000 (0:00:01.935)       0:01:21.892 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:03:59 +0000 (0:00:01.194)       0:01:23.086 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:03:52.185577', 'end': '2026-03-26 05:03:52.243863', 'delta': '0:00:00.058286', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:03:53.038758', 'end': '2026-03-26 05:03:53.088240', 'delta': '0:00:00.049482', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:03:53.633676', 'end': '2026-03-26 05:03:53.688108', 'delta': '0:00:00.054432', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:04:00 +0000 (0:00:01.242)       0:01:24.329 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:04:02 +0000 (0:00:02.174)       0:01:26.504 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:04:04 +0000 (0:00:01.279)       0:01:27.784 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:04:06 +0000 (0:00:02.152)       0:01:29.937 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:04:09 +0000 (0:00:03.367)       0:01:33.305 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:04:11 +0000 (0:00:02.171)       0:01:35.477 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:04:12 +0000 (0:00:01.175)       0:01:36.652 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:04:14 +0000 (0:00:01.207)       0:01:37.859 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:04:16 +0000 (0:00:02.518)       0:01:40.378 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:04:18 +0000 (0:00:01.888)       0:01:42.266 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:04:20 +0000 (0:00:02.085)       0:01:44.352 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:04:22 +0000 (0:00:01.949)       0:01:46.301 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:04:24 +0000 (0:00:02.102)       0:01:48.404 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:04:26 +0000 (0:00:02.017)       0:01:50.421 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:04:28 +0000 (0:00:02.134)       0:01:52.556 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-40-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1060', 'sectorsize': '2048', 'size': '530.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2ece1d7d', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:04:31 +0000 (0:00:02.526)       0:01:55.083 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-40-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1060', 'sectorsize': '2048', 'size': '530.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2ece1d7d', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:04:33 +0000 (0:00:02.390)       0:01:57.474 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:04:36 +0000 (0:00:02.554)       0:02:00.029 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:04:38 +0000 (0:00:02.027)       0:02:02.056 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:04:40 +0000 (0:00:02.483)       0:02:04.540 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:04:42 +0000 (0:00:01.950)       0:02:06.491 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:04:45 +0000 (0:00:02.589)       0:02:09.080 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:04:47 +0000 (0:00:01.980)       0:02:11.061 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:04:51 +0000 (0:00:04.227)       0:02:15.289 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:04:53 +0000 (0:00:02.297)       0:02:17.586 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:04:55 +0000 (0:00:02.018)       0:02:19.605 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:04:57 +0000 (0:00:01.688)       0:02:21.294 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:04:59 +0000 (0:00:01.410)       0:02:22.704 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:05:00 +0000 (0:00:01.340)       0:02:24.045 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:05:01 +0000 (0:00:01.414)       0:02:25.460 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:05:03 +0000 (0:00:01.677)       0:02:27.138 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:05:05 +0000 (0:00:01.718)       0:02:28.856 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:05:06 +0000 (0:00:01.586)       0:02:30.443 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:05:08 +0000 (0:00:01.430)       0:02:31.874 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:05:09 +0000 (0:00:01.498)       0:02:33.372 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:05:11 +0000 (0:00:01.970)       0:02:35.343 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-infra : Update cache for Debian based OSs] **************************\nThursday 26 March 2026  05:05:14 +0000 (0:00:03.036)       0:02:38.380 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-infra : Include_tasks configure_firewall.yml] ***********************\nThursday 26 March 2026  05:05:22 +0000 (0:00:08.251)       0:02:46.631 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-infra : Include_tasks setup_ntp.yml] ********************************\nThursday 26 March 2026  05:05:25 +0000 (0:00:02.179)       0:02:48.811 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-infra : Add logrotate configuration] ********************************\nThursday 26 March 2026  05:05:27 +0000 (0:00:01.911)       0:02:50.723 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Include check_system.yml] ********************************\nThursday 26 March 2026  05:05:30 +0000 (0:00:03.105)       0:02:53.828 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_system.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported ansible version (1.X)] ***************\nThursday 26 March 2026  05:05:33 +0000 (0:00:02.913)       0:02:56.741 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported system] ******************************\nThursday 26 March 2026  05:05:34 +0000 (0:00:01.808)       0:02:58.550 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported architecture] ************************\nThursday 26 March 2026  05:05:37 +0000 (0:00:02.111)       0:03:00.662 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported distribution] ************************\nThursday 26 March 2026  05:05:39 +0000 (0:00:02.036)       0:03:02.698 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported CentOS release] **********************\nThursday 26 March 2026  05:05:41 +0000 (0:00:02.169)       0:03:04.867 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported distribution for ubuntu cloud archive] ***\nThursday 26 March 2026  05:05:43 +0000 (0:00:01.955)       0:03:06.823 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail on unsupported SUSE/openSUSE distribution (only 15.x supported)] ***\nThursday 26 March 2026  05:05:45 +0000 (0:00:02.173)       0:03:08.997 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail if systemd is not present] **************************\nThursday 26 March 2026  05:05:47 +0000 (0:00:01.947)       0:03:10.944 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate repository variables in non-containerized scenario] ***\nThursday 26 March 2026  05:05:49 +0000 (0:00:02.267)       0:03:13.212 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate osd_objectstore] ********************************\nThursday 26 March 2026  05:05:51 +0000 (0:00:02.082)       0:03:15.295 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate radosgw network configuration] ******************\nThursday 26 March 2026  05:05:53 +0000 (0:00:02.175)       0:03:17.470 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate lvm osd scenario] *******************************\nThursday 26 March 2026  05:05:56 +0000 (0:00:02.326)       0:03:19.796 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate bluestore lvm osd scenario] *********************\nThursday 26 March 2026  05:05:58 +0000 (0:00:01.940)       0:03:21.737 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail if local scenario is enabled on debian] *************\nThursday 26 March 2026  05:06:00 +0000 (0:00:02.288)       0:03:24.025 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail if rhcs repository is enabled on debian] ************\nThursday 26 March 2026  05:06:02 +0000 (0:00:01.980)       0:03:26.005 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Check ceph_origin definition on SUSE/openSUSE Leap] ******\nThursday 26 March 2026  05:06:04 +0000 (0:00:02.178)       0:03:28.184 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Check ceph_repository definition on SUSE/openSUSE Leap] ***\nThursday 26 March 2026  05:06:06 +0000 (0:00:01.996)       0:03:30.181 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate ntp daemon type] ********************************\nThursday 26 March 2026  05:06:08 +0000 (0:00:02.413)       0:03:32.594 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Abort if ntp_daemon_type is ntpd on Atomic] **************\nThursday 26 March 2026  05:06:10 +0000 (0:00:02.030)       0:03:34.625 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Include check_devices.yml] *******************************\nThursday 26 March 2026  05:06:13 +0000 (0:00:02.077)       0:03:36.703 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_devices.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-validate : Set_fact root_device] ************************************\nThursday 26 March 2026  05:06:15 +0000 (0:00:02.515)       0:03:39.218 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Resolve devices in lvm_volumes] **************************\nThursday 26 March 2026  05:06:17 +0000 (0:00:01.463)       0:03:40.681 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Set_fact lvm_volumes_data_devices] ***********************\nThursday 26 March 2026  05:06:18 +0000 (0:00:01.382)       0:03:42.064 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if root_device is passed in lvm_volumes or devices] ***\nThursday 26 March 2026  05:06:20 +0000 (0:00:01.660)       0:03:43.725 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Get devices information] *********************************\nThursday 26 March 2026  05:06:21 +0000 (0:00:01.449)       0:03:45.175 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if one of the devices is not a device] **************\nThursday 26 March 2026  05:06:22 +0000 (0:00:01.356)       0:03:46.531 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail when gpt header found on osd devices] ***************\nThursday 26 March 2026  05:06:24 +0000 (0:00:01.413)       0:03:47.944 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Check data logical volume] *******************************\nThursday 26 March 2026  05:06:25 +0000 (0:00:01.325)       0:03:49.270 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m\n\nTASK [ceph-validate : Fail if one of the data logical volume is not a device or doesn't exist] ***\nThursday 26 March 2026  05:06:28 +0000 (0:00:02.964)       0:03:52.234 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a/osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 955, 'dev': 6, 'nlink': 1, 'atime': 1774493776.0892901, 'mtime': 1774493776.08229, 'ctime': 1774493776.08229, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a/osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-e2623153-bc41-510f-8884-ef957bb96082/osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 965, 'dev': 6, 'nlink': 1, 'atime': 1774493795.0405843, 'mtime': 1774493795.0355842, 'ctime': 1774493795.0355842, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-e2623153-bc41-510f-8884-ef957bb96082/osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-a652979e-9f40-503a-bbc8-6de5e605991e/osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 957, 'dev': 6, 'nlink': 1, 'atime': 1774493778.800435, 'mtime': 1774493778.795435, 'ctime': 1774493778.795435, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-a652979e-9f40-503a-bbc8-6de5e605991e/osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-b5eee7c3-8883-5bbe-be5a-75726e822543/osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 967, 'dev': 6, 'nlink': 1, 'atime': 1774493797.1287172, 'mtime': 1774493797.1227171, 'ctime': 1774493797.1227171, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-b5eee7c3-8883-5bbe-be5a-75726e822543/osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66/osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 949, 'dev': 6, 'nlink': 1, 'atime': 1774493778.807298, 'mtime': 1774493778.8022978, 'ctime': 1774493778.8022978, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66/osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771/osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 959, 'dev': 6, 'nlink': 1, 'atime': 1774493796.95358, 'mtime': 1774493796.9495797, 'ctime': 1774493796.9495797, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771/osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Check bluestore db logical volume] ***********************\nThursday 26 March 2026  05:06:30 +0000 (0:00:01.595)       0:03:53.830 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if one of the bluestore db logical volume is not a device or doesn't exist] ***\nThursday 26 March 2026  05:06:31 +0000 (0:00:01.451)       0:03:55.281 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Check bluestore wal logical volume] **********************\nThursday 26 March 2026  05:06:33 +0000 (0:00:01.497)       0:03:56.779 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if one of the bluestore wal logical volume is not a device or doesn't exist] ***\nThursday 26 March 2026  05:06:34 +0000 (0:00:01.663)       0:03:58.443 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Include check_eth_rgw.yml] *******************************\nThursday 26 March 2026  05:06:36 +0000 (0:00:01.384)       0:03:59.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Include check_rgw_pools.yml] *****************************\nThursday 26 March 2026  05:06:38 +0000 (0:00:01.971)       0:04:01.800 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_rgw_pools.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-validate : Fail if ec_profile is not set for ec pools] **************\nThursday 26 March 2026  05:06:40 +0000 (0:00:02.635)       0:04:04.435 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if ec_k is not set for ec pools] ********************\nThursday 26 March 2026  05:06:42 +0000 (0:00:01.442)       0:04:05.877 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Fail if ec_m is not set for ec pools] ********************\nThursday 26 March 2026  05:06:43 +0000 (0:00:01.713)       0:04:07.591 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Include check_nfs.yml] ***********************************\nThursday 26 March 2026  05:06:45 +0000 (0:00:01.509)       0:04:09.100 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Include check_rbdmirror.yml] *****************************\nThursday 26 March 2026  05:06:47 +0000 (0:00:01.861)       0:04:10.961 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail if monitoring group doesn't exist] ******************\nThursday 26 March 2026  05:06:49 +0000 (0:00:02.202)       0:04:13.164 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail when monitoring doesn't contain at least one node.] ***\nThursday 26 March 2026  05:06:51 +0000 (0:00:02.090)       0:04:15.254 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Fail when dashboard_admin_password and/or grafana_admin_password are not set] ***\nThursday 26 March 2026  05:06:53 +0000 (0:00:02.016)       0:04:17.271 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate container registry credentials] *****************\nThursday 26 March 2026  05:06:55 +0000 (0:00:02.121)       0:04:19.393 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate container service and container package] ********\nThursday 26 March 2026  05:06:57 +0000 (0:00:02.059)       0:04:21.452 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate openstack_keys key format] **********************\nThursday 26 March 2026  05:07:00 +0000 (0:00:02.271)       0:04:23.724 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Validate clients keys key format] ************************\nThursday 26 March 2026  05:07:02 +0000 (0:00:02.175)       0:04:25.899 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate openstack_keys caps] ****************************\nThursday 26 March 2026  05:07:04 +0000 (0:00:02.159)       0:04:28.058 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-validate : Validate clients keys caps] ******************************\nThursday 26 March 2026  05:07:06 +0000 (0:00:02.193)       0:04:30.252 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Check virtual_ips is defined] ****************************\nThursday 26 March 2026  05:07:08 +0000 (0:00:02.309)       0:04:32.562 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-validate : Validate virtual_ips length] *****************************\nThursday 26 March 2026  05:07:10 +0000 (0:00:02.062)       0:04:34.624 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-engine : Include pre_requisites/prerequisites.yml] ********\nThursday 26 March 2026  05:07:13 +0000 (0:00:02.258)       0:04:36.883 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m\n\nTASK [ceph-container-engine : Include specific variables] **********************\nThursday 26 March 2026  05:07:16 +0000 (0:00:02.867)       0:04:39.750 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m\n\nTASK [ceph-container-engine : Create the systemd docker override directory] ****\nThursday 26 March 2026  05:07:18 +0000 (0:00:02.053)       0:04:41.804 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-engine : Create the systemd docker override file] *********\nThursday 26 March 2026  05:07:20 +0000 (0:00:02.167)       0:04:43.971 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-engine : Remove docker proxy configuration] ***************\nThursday 26 March 2026  05:07:22 +0000 (0:00:01.993)       0:04:45.965 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-engine : Restart docker] **********************************\nThursday 26 March 2026  05:07:24 +0000 (0:00:02.604)       0:04:48.570 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-common : Container registry authentication] ***************\nThursday 26 March 2026  05:07:27 +0000 (0:00:02.420)       0:04:50.990 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [Get the ceph release being deployed] *************************************\nThursday 26 March 2026  05:07:29 +0000 (0:00:02.355)       0:04:53.345 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Check ceph release being deployed] ***************************************\nThursday 26 March 2026  05:07:32 +0000 (0:00:02.637)       0:04:55.982 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Ensure cluster config is applied] ****************************************\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:07:33 +0000 (0:00:01.466)       0:04:57.450 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:07:35 +0000 (0:00:01.455)       0:04:58.905 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set cluster configs] *****************************************************\nThursday 26 March 2026  05:07:36 +0000 (0:00:01.150)       0:05:00.056 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m\n\nPLAY [Upgrade ceph mon cluster] ************************************************\n\nTASK [Remove ceph aliases] *****************************************************\nThursday 26 March 2026  05:07:46 +0000 (0:00:10.316)       0:05:10.373 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set mon_host_count] ******************************************************\nThursday 26 March 2026  05:07:48 +0000 (0:00:01.489)       0:05:11.862 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail when less than three monitors] **************************************\nThursday 26 March 2026  05:07:49 +0000 (0:00:01.151)       0:05:13.014 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Select a running monitor] ************************************************\nThursday 26 March 2026  05:07:50 +0000 (0:00:01.210)       0:05:14.224 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:07:51 +0000 (0:00:01.156)       0:05:15.381 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:07:52 +0000 (0:00:01.184)       0:05:16.566 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:07:54 +0000 (0:00:01.428)       0:05:17.994 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:07:55 +0000 (0:00:01.122)       0:05:19.117 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:07:56 +0000 (0:00:01.409)       0:05:20.527 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:07:58 +0000 (0:00:01.204)       0:05:21.732 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:07:59 +0000 (0:00:01.173)       0:05:22.905 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:08:00 +0000 (0:00:01.276)       0:05:24.181 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:08:01 +0000 (0:00:01.114)       0:05:25.296 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:08:02 +0000 (0:00:01.193)       0:05:26.489 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:08:04 +0000 (0:00:01.716)       0:05:28.206 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:08:05 +0000 (0:00:01.228)       0:05:29.434 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:08:09 +0000 (0:00:03.287)       0:05:32.722 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:08:10 +0000 (0:00:01.380)       0:05:34.102 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:08:12 +0000 (0:00:01.865)       0:05:35.968 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:08:13 +0000 (0:00:01.175)       0:05:37.144 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:08:06.328851', 'end': '2026-03-26 05:08:06.373241', 'delta': '0:00:00.044390', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:08:06.897478', 'end': '2026-03-26 05:08:06.954848', 'delta': '0:00:00.057370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:08:07.774569', 'end': '2026-03-26 05:08:07.828855', 'delta': '0:00:00.054286', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:08:14 +0000 (0:00:01.215)       0:05:38.360 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:08:16 +0000 (0:00:01.615)       0:05:39.975 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:08:17 +0000 (0:00:01.231)       0:05:41.207 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:08:18 +0000 (0:00:01.138)       0:05:42.346 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:08:20 +0000 (0:00:02.054)       0:05:44.401 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:08:21 +0000 (0:00:01.198)       0:05:45.599 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:08:23 +0000 (0:00:01.203)       0:05:46.803 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:08:24 +0000 (0:00:01.285)       0:05:48.088 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:08:25 +0000 (0:00:01.162)       0:05:49.250 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:08:26 +0000 (0:00:01.126)       0:05:50.377 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:08:27 +0000 (0:00:01.134)       0:05:51.512 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:08:28 +0000 (0:00:01.116)       0:05:52.628 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:08:30 +0000 (0:00:01.133)       0:05:53.762 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:08:31 +0000 (0:00:01.128)       0:05:54.891 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:08:32 +0000 (0:00:01.139)       0:05:56.031 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:08:33 +0000 (0:00:01.223)       0:05:57.254 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:08:34 +0000 (0:00:01.245)       0:05:58.500 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:08:36 +0000 (0:00:01.553)       0:06:00.054 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:08:37 +0000 (0:00:01.144)       0:06:01.199 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:08:39 +0000 (0:00:01.521)       0:06:02.720 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:08:40 +0000 (0:00:01.103)       0:06:03.824 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:08:41 +0000 (0:00:01.244)       0:06:05.068 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:08:42 +0000 (0:00:01.198)       0:06:06.267 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:08:44 +0000 (0:00:01.994)       0:06:08.262 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:08:45 +0000 (0:00:01.173)       0:06:09.435 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:08:46 +0000 (0:00:01.145)       0:06:10.581 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:08:49 +0000 (0:00:02.201)       0:06:12.782 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get ceph cluster status] *************************************************\nThursday 26 March 2026  05:08:52 +0000 (0:00:02.906)       0:06:15.689 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [Display ceph health detail] **********************************************\nThursday 26 March 2026  05:08:54 +0000 (0:00:02.285)       0:06:17.974 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail if cluster isn't in an acceptable state] ****************************\nThursday 26 March 2026  05:08:55 +0000 (0:00:01.227)       0:06:19.202 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Get the ceph quorum status] **********************************************\nThursday 26 March 2026  05:08:56 +0000 (0:00:01.140)       0:06:20.342 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [Fail if the cluster quorum isn't in an acceptable state] *****************\nThursday 26 March 2026  05:08:58 +0000 (0:00:02.264)       0:06:22.607 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************\nThursday 26 March 2026  05:09:00 +0000 (0:00:01.162)       0:06:23.769 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [Create potentially missing keys (rbd and rbd-mirror)] ********************\nThursday 26 March 2026  05:09:02 +0000 (0:00:02.576)       0:06:26.346 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=['bootstrap-rbd', 'testbed-node-0'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=['bootstrap-rbd', 'testbed-node-1'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=['bootstrap-rbd', 'testbed-node-2'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=['bootstrap-rbd-mirror', 'testbed-node-0'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=['bootstrap-rbd-mirror', 'testbed-node-1'])\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=['bootstrap-rbd-mirror', 'testbed-node-2'])\u001b[0m\n\nTASK [Stop ceph mon] ***********************************************************\nThursday 26 March 2026  05:09:16 +0000 (0:00:13.506)       0:06:39.853 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [Mask the mgr service] ****************************************************\nThursday 26 March 2026  05:09:19 +0000 (0:00:03.753)       0:06:43.607 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:09:22 +0000 (0:00:02.565)       0:06:46.173 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:09:24 +0000 (0:00:01.549)       0:06:47.723 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:09:25 +0000 (0:00:01.598)       0:06:49.322 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:09:27 +0000 (0:00:01.528)       0:06:50.850 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:09:28 +0000 (0:00:01.109)       0:06:51.960 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:09:29 +0000 (0:00:01.111)       0:06:53.071 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:09:30 +0000 (0:00:01.150)       0:06:54.222 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:09:32 +0000 (0:00:01.605)       0:06:55.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:09:33 +0000 (0:00:01.146)       0:06:56.975 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:09:34 +0000 (0:00:01.103)       0:06:58.078 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:09:36 +0000 (0:00:01.650)       0:06:59.729 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:09:37 +0000 (0:00:01.553)       0:07:01.282 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:09:38 +0000 (0:00:01.188)       0:07:02.471 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:09:39 +0000 (0:00:01.181)       0:07:03.653 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:09:41 +0000 (0:00:01.153)       0:07:04.807 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:09:42 +0000 (0:00:01.111)       0:07:05.918 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:09:43 +0000 (0:00:01.152)       0:07:07.070 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:09:44 +0000 (0:00:01.133)       0:07:08.204 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:09:45 +0000 (0:00:01.162)       0:07:09.366 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:09:46 +0000 (0:00:01.121)       0:07:10.488 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:09:47 +0000 (0:00:01.163)       0:07:11.652 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:09:49 +0000 (0:00:01.175)       0:07:12.828 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:09:50 +0000 (0:00:01.104)       0:07:13.933 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:09:51 +0000 (0:00:01.120)       0:07:15.053 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:09:52 +0000 (0:00:01.141)       0:07:16.195 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:09:53 +0000 (0:00:01.173)       0:07:17.369 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:09:54 +0000 (0:00:01.128)       0:07:18.497 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:09:56 +0000 (0:00:01.174)       0:07:19.671 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:09:57 +0000 (0:00:01.108)       0:07:20.779 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:09:58 +0000 (0:00:01.130)       0:07:21.909 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:09:59 +0000 (0:00:01.129)       0:07:23.038 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:10:00 +0000 (0:00:01.211)       0:07:24.250 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:10:01 +0000 (0:00:01.136)       0:07:25.386 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:10:02 +0000 (0:00:01.153)       0:07:26.540 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:10:04 +0000 (0:00:02.009)       0:07:28.549 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:10:07 +0000 (0:00:02.412)       0:07:30.962 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:10:08 +0000 (0:00:01.491)       0:07:32.454 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:10:09 +0000 (0:00:01.135)       0:07:33.590 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:10:11 +0000 (0:00:01.150)       0:07:34.740 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:10:13 +0000 (0:00:02.012)       0:07:36.752 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:10:14 +0000 (0:00:01.644)       0:07:38.396 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:10:15 +0000 (0:00:01.178)       0:07:39.575 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:10:17 +0000 (0:00:01.122)       0:07:40.698 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:10:18 +0000 (0:00:01.105)       0:07:41.803 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:10:19 +0000 (0:00:01.442)       0:07:43.245 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:10:21 +0000 (0:00:01.919)       0:07:45.165 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:10:22 +0000 (0:00:01.207)       0:07:46.373 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:10:23 +0000 (0:00:01.133)       0:07:47.506 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:10:25 +0000 (0:00:01.168)       0:07:48.674 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:10:26 +0000 (0:00:01.181)       0:07:49.856 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:10:27 +0000 (0:00:01.145)       0:07:51.001 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:10:28 +0000 (0:00:01.146)       0:07:52.148 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:10:31 +0000 (0:00:02.586)       0:07:54.734 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:10:32 +0000 (0:00:01.153)       0:07:55.888 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:10:33 +0000 (0:00:01.486)       0:07:57.375 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:10:34 +0000 (0:00:01.192)       0:07:58.568 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:10:36 +0000 (0:00:01.141)       0:07:59.710 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:10:37 +0000 (0:00:01.119)       0:08:00.829 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:10:38 +0000 (0:00:01.121)       0:08:01.951 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:10:39 +0000 (0:00:01.190)       0:08:03.142 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:10:40 +0000 (0:00:01.162)       0:08:04.304 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:10:41 +0000 (0:00:01.117)       0:08:05.422 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:10:42 +0000 (0:00:01.169)       0:08:06.591 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:10:44 +0000 (0:00:01.150)       0:08:07.741 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:10:45 +0000 (0:00:01.555)       0:08:09.297 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:10:52 +0000 (0:00:06.764)       0:08:16.062 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:10:53 +0000 (0:00:01.113)       0:08:17.175 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:10:54 +0000 (0:00:01.125)       0:08:18.301 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:10:55 +0000 (0:00:01.132)       0:08:19.434 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:10:56 +0000 (0:00:01.146)       0:08:20.581 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:10:58 +0000 (0:00:01.110)       0:08:21.691 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:10:59 +0000 (0:00:01.194)       0:08:22.885 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:11:00 +0000 (0:00:01.126)       0:08:24.012 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:11:01 +0000 (0:00:01.116)       0:08:25.128 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:11:02 +0000 (0:00:01.154)       0:08:26.283 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:11:03 +0000 (0:00:01.115)       0:08:27.399 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:11:04 +0000 (0:00:01.158)       0:08:28.558 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:11:06 +0000 (0:00:01.164)       0:08:29.722 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:11:07 +0000 (0:00:01.257)       0:08:30.980 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:11:08 +0000 (0:00:01.165)       0:08:32.147 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:11:09 +0000 (0:00:01.213)       0:08:33.360 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:11:10 +0000 (0:00:01.193)       0:08:34.554 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:11:12 +0000 (0:00:01.190)       0:08:35.744 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:11:13 +0000 (0:00:01.119)       0:08:36.864 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:11:14 +0000 (0:00:01.194)       0:08:38.059 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:11:15 +0000 (0:00:01.131)       0:08:39.190 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:11:16 +0000 (0:00:01.137)       0:08:40.328 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:11:18 +0000 (0:00:01.768)       0:08:42.096 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:11:19 +0000 (0:00:01.440)       0:08:43.537 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:11:21 +0000 (0:00:01.477)       0:08:45.015 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:11:22 +0000 (0:00:01.150)       0:08:46.166 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:11:23 +0000 (0:00:01.405)       0:08:47.571 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:11:25 +0000 (0:00:01.819)       0:08:49.391 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nThursday 26 March 2026  05:11:26 +0000 (0:00:01.148)       0:08:50.540 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nThursday 26 March 2026  05:11:28 +0000 (0:00:01.527)       0:08:52.068 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nThursday 26 March 2026  05:11:31 +0000 (0:00:03.507)       0:08:55.575 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nThursday 26 March 2026  05:11:33 +0000 (0:00:01.206)       0:08:56.781 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nThursday 26 March 2026  05:11:34 +0000 (0:00:01.146)       0:08:57.928 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nThursday 26 March 2026  05:11:35 +0000 (0:00:01.166)       0:08:59.095 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nThursday 26 March 2026  05:11:37 +0000 (0:00:02.085)       0:09:01.180 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nThursday 26 March 2026  05:11:39 +0000 (0:00:01.613)       0:09:02.794 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nThursday 26 March 2026  05:11:40 +0000 (0:00:01.465)       0:09:04.260 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nThursday 26 March 2026  05:11:42 +0000 (0:00:01.542)       0:09:05.802 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nThursday 26 March 2026  05:11:43 +0000 (0:00:01.688)       0:09:07.491 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nThursday 26 March 2026  05:11:45 +0000 (0:00:01.696)       0:09:09.188 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nThursday 26 March 2026  05:11:49 +0000 (0:00:03.888)       0:09:13.076 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nThursday 26 March 2026  05:11:51 +0000 (0:00:02.132)       0:09:15.209 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nThursday 26 March 2026  05:11:52 +0000 (0:00:01.155)       0:09:16.365 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nThursday 26 March 2026  05:11:53 +0000 (0:00:01.179)       0:09:17.544 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nThursday 26 March 2026  05:11:55 +0000 (0:00:02.047)       0:09:19.591 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nThursday 26 March 2026  05:11:57 +0000 (0:00:01.416)       0:09:21.008 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nThursday 26 March 2026  05:11:58 +0000 (0:00:01.144)       0:09:22.152 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:11:59 +0000 (0:00:01.472)       0:09:23.624 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nThursday 26 March 2026  05:12:01 +0000 (0:00:01.125)       0:09:24.750 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:12:02 +0000 (0:00:01.105)       0:09:25.855 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nThursday 26 March 2026  05:12:03 +0000 (0:00:01.470)       0:09:27.326 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nThursday 26 March 2026  05:12:05 +0000 (0:00:02.330)       0:09:29.657 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nThursday 26 March 2026  05:12:08 +0000 (0:00:02.018)       0:09:31.675 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nThursday 26 March 2026  05:12:10 +0000 (0:00:02.419)       0:09:34.095 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nThursday 26 March 2026  05:12:13 +0000 (0:00:03.290)       0:09:37.386 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nThursday 26 March 2026  05:12:15 +0000 (0:00:01.623)       0:09:39.010 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nThursday 26 March 2026  05:12:17 +0000 (0:00:02.355)       0:09:41.365 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nThursday 26 March 2026  05:12:20 +0000 (0:00:02.967)       0:09:44.332 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nThursday 26 March 2026  05:12:21 +0000 (0:00:01.131)       0:09:45.464 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m\n\nTASK [Start ceph mgr] **********************************************************\nThursday 26 March 2026  05:12:31 +0000 (0:00:09.703)       0:09:55.167 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:12:33 +0000 (0:00:02.422)       0:09:57.590 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:12:36 +0000 (0:00:02.229)       0:09:59.820 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Non container | waiting for the monitor to join the quorum...] ***********\nThursday 26 March 2026  05:12:37 +0000 (0:00:01.430)       0:10:01.251 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Container | waiting for the containerized monitor to join the quorum...] ***\nThursday 26 March 2026  05:12:38 +0000 (0:00:01.139)       0:10:02.391 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:12:40 +0000 (0:00:02.224)       0:10:04.615 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:12:42 +0000 (0:00:01.124)       0:10:05.740 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:12:43 +0000 (0:00:01.100)       0:10:06.841 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:12:44 +0000 (0:00:01.117)       0:10:07.958 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:12:45 +0000 (0:00:01.131)       0:10:09.089 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:12:46 +0000 (0:00:01.091)       0:10:10.181 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:12:47 +0000 (0:00:01.121)       0:10:11.303 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Upgrade ceph mon cluster] ************************************************\n\nTASK [Remove ceph aliases] *****************************************************\nThursday 26 March 2026  05:12:48 +0000 (0:00:00.939)       0:10:12.242 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Set mon_host_count] ******************************************************\nThursday 26 March 2026  05:12:49 +0000 (0:00:01.134)       0:10:13.377 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Fail when less than three monitors] **************************************\nThursday 26 March 2026  05:12:50 +0000 (0:00:00.796)       0:10:14.173 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Select a running monitor] ************************************************\nThursday 26 March 2026  05:12:51 +0000 (0:00:00.751)       0:10:14.925 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:12:52 +0000 (0:00:00.774)       0:10:15.699 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:12:53 +0000 (0:00:01.197)       0:10:16.897 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:12:54 +0000 (0:00:01.482)       0:10:18.379 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:12:55 +0000 (0:00:01.155)       0:10:19.535 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:12:57 +0000 (0:00:01.444)       0:10:20.979 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:12:58 +0000 (0:00:01.141)       0:10:22.120 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:12:59 +0000 (0:00:01.146)       0:10:23.267 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:13:00 +0000 (0:00:01.138)       0:10:24.406 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:13:01 +0000 (0:00:01.137)       0:10:25.544 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:13:03 +0000 (0:00:01.225)       0:10:26.769 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:13:05 +0000 (0:00:01.979)       0:10:28.749 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:13:06 +0000 (0:00:01.291)       0:10:30.040 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:13:09 +0000 (0:00:03.220)       0:10:33.261 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:13:11 +0000 (0:00:01.731)       0:10:34.993 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:13:13 +0000 (0:00:01.960)       0:10:36.954 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:13:14 +0000 (0:00:01.173)       0:10:38.127 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:13:06.930490', 'end': '2026-03-26 05:13:06.985292', 'delta': '0:00:00.054802', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:13:07.808352', 'end': '2026-03-26 05:13:07.863473', 'delta': '0:00:00.055121', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:13:08.378617', 'end': '2026-03-26 05:13:08.427807', 'delta': '0:00:00.049190', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:13:15 +0000 (0:00:01.248)       0:10:39.375 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:13:16 +0000 (0:00:01.265)       0:10:40.640 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:13:18 +0000 (0:00:01.289)       0:10:41.930 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:13:19 +0000 (0:00:01.171)       0:10:43.101 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:13:22 +0000 (0:00:02.994)       0:10:46.095 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:13:23 +0000 (0:00:01.136)       0:10:47.232 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:13:24 +0000 (0:00:01.163)       0:10:48.395 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:13:25 +0000 (0:00:01.199)       0:10:49.595 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:13:27 +0000 (0:00:01.130)       0:10:50.725 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:13:28 +0000 (0:00:01.148)       0:10:51.874 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:13:29 +0000 (0:00:01.191)       0:10:53.065 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:13:30 +0000 (0:00:01.161)       0:10:54.227 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:13:31 +0000 (0:00:01.180)       0:10:55.408 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:13:32 +0000 (0:00:01.116)       0:10:56.525 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:13:33 +0000 (0:00:01.105)       0:10:57.630 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:13:35 +0000 (0:00:01.240)       0:10:58.870 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:13:36 +0000 (0:00:01.281)       0:11:00.152 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:13:38 +0000 (0:00:01.512)       0:11:01.665 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:13:39 +0000 (0:00:01.166)       0:11:02.831 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:13:40 +0000 (0:00:01.499)       0:11:04.331 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:13:41 +0000 (0:00:01.115)       0:11:05.447 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:13:43 +0000 (0:00:01.408)       0:11:06.855 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:13:44 +0000 (0:00:01.164)       0:11:08.019 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:13:46 +0000 (0:00:02.048)       0:11:10.068 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:13:47 +0000 (0:00:01.208)       0:11:11.277 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:13:48 +0000 (0:00:01.111)       0:11:12.389 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:13:50 +0000 (0:00:01.883)       0:11:14.272 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get ceph cluster status] *************************************************\nThursday 26 March 2026  05:13:53 +0000 (0:00:02.391)       0:11:16.664 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Display ceph health detail] **********************************************\nThursday 26 March 2026  05:13:53 +0000 (0:00:00.872)       0:11:17.537 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Fail if cluster isn't in an acceptable state] ****************************\nThursday 26 March 2026  05:13:54 +0000 (0:00:00.876)       0:11:18.414 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Get the ceph quorum status] **********************************************\nThursday 26 March 2026  05:13:55 +0000 (0:00:00.765)       0:11:19.180 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Fail if the cluster quorum isn't in an acceptable state] *****************\nThursday 26 March 2026  05:13:56 +0000 (0:00:00.894)       0:11:20.074 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************\nThursday 26 March 2026  05:13:57 +0000 (0:00:00.785)       0:11:20.859 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Create potentially missing keys (rbd and rbd-mirror)] ********************\nThursday 26 March 2026  05:13:58 +0000 (0:00:01.087)       0:11:21.947 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-1']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-2']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-1']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-2']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Stop ceph mon] ***********************************************************\nThursday 26 March 2026  05:13:59 +0000 (0:00:01.605)       0:11:23.552 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\nTASK [Mask the mgr service] ****************************************************\nThursday 26 March 2026  05:14:03 +0000 (0:00:03.184)       0:11:26.737 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:14:05 +0000 (0:00:01.984)       0:11:28.721 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:14:06 +0000 (0:00:01.137)       0:11:29.859 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:14:07 +0000 (0:00:01.138)       0:11:30.998 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:14:08 +0000 (0:00:01.518)       0:11:32.517 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:14:09 +0000 (0:00:01.128)       0:11:33.645 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:14:11 +0000 (0:00:01.116)       0:11:34.762 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:14:12 +0000 (0:00:01.238)       0:11:36.000 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:14:13 +0000 (0:00:01.551)       0:11:37.552 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:14:15 +0000 (0:00:01.151)       0:11:38.704 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:14:16 +0000 (0:00:01.162)       0:11:39.866 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:14:17 +0000 (0:00:01.529)       0:11:41.396 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:14:19 +0000 (0:00:01.555)       0:11:42.951 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:14:20 +0000 (0:00:00.806)       0:11:43.758 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:14:20 +0000 (0:00:00.824)       0:11:44.582 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:14:21 +0000 (0:00:00.801)       0:11:45.383 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:14:22 +0000 (0:00:00.798)       0:11:46.182 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:14:23 +0000 (0:00:00.775)       0:11:46.957 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:14:24 +0000 (0:00:00.805)       0:11:47.762 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:14:24 +0000 (0:00:00.819)       0:11:48.582 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:14:25 +0000 (0:00:00.865)       0:11:49.448 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:14:26 +0000 (0:00:00.805)       0:11:50.253 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:14:27 +0000 (0:00:00.831)       0:11:51.085 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:14:28 +0000 (0:00:00.766)       0:11:51.852 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:14:28 +0000 (0:00:00.789)       0:11:52.642 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:14:29 +0000 (0:00:00.879)       0:11:53.521 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:14:30 +0000 (0:00:00.760)       0:11:54.281 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:14:31 +0000 (0:00:00.787)       0:11:55.068 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:14:32 +0000 (0:00:00.783)       0:11:55.852 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:14:32 +0000 (0:00:00.782)       0:11:56.634 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:14:33 +0000 (0:00:00.824)       0:11:57.459 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:14:34 +0000 (0:00:00.765)       0:11:58.224 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:14:35 +0000 (0:00:00.833)       0:11:59.058 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:14:36 +0000 (0:00:00.780)       0:11:59.839 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:14:37 +0000 (0:00:00.826)       0:12:00.665 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:14:38 +0000 (0:00:01.617)       0:12:02.283 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:14:40 +0000 (0:00:02.084)       0:12:04.367 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:14:41 +0000 (0:00:01.209)       0:12:05.576 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:14:43 +0000 (0:00:01.134)       0:12:06.711 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:14:44 +0000 (0:00:01.146)       0:12:07.858 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:14:46 +0000 (0:00:01.852)       0:12:09.711 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:14:47 +0000 (0:00:01.537)       0:12:11.249 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:14:48 +0000 (0:00:01.162)       0:12:12.411 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:14:49 +0000 (0:00:00.805)       0:12:13.216 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:14:50 +0000 (0:00:00.755)       0:12:13.972 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:14:51 +0000 (0:00:01.138)       0:12:15.111 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:14:53 +0000 (0:00:01.856)       0:12:16.967 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:14:54 +0000 (0:00:01.130)       0:12:18.098 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:14:55 +0000 (0:00:01.155)       0:12:19.253 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:14:56 +0000 (0:00:01.211)       0:12:20.464 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:14:58 +0000 (0:00:01.200)       0:12:21.665 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:14:59 +0000 (0:00:01.192)       0:12:22.857 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:15:00 +0000 (0:00:00.811)       0:12:23.669 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:15:02 +0000 (0:00:02.213)       0:12:25.883 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:15:03 +0000 (0:00:00.787)       0:12:26.671 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:15:04 +0000 (0:00:01.115)       0:12:27.787 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:15:05 +0000 (0:00:01.188)       0:12:28.975 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:15:06 +0000 (0:00:01.138)       0:12:30.114 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:15:07 +0000 (0:00:01.115)       0:12:31.230 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:15:08 +0000 (0:00:01.155)       0:12:32.385 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:15:09 +0000 (0:00:01.131)       0:12:33.516 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:15:10 +0000 (0:00:01.140)       0:12:34.656 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:15:12 +0000 (0:00:01.206)       0:12:35.862 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:15:13 +0000 (0:00:01.119)       0:12:36.982 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:15:14 +0000 (0:00:00.790)       0:12:37.773 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:15:15 +0000 (0:00:01.126)       0:12:38.899 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:15:21 +0000 (0:00:06.237)       0:12:45.137 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:15:22 +0000 (0:00:00.794)       0:12:45.931 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:15:23 +0000 (0:00:00.780)       0:12:46.711 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:15:23 +0000 (0:00:00.787)       0:12:47.498 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:15:24 +0000 (0:00:00.796)       0:12:48.295 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:15:25 +0000 (0:00:00.840)       0:12:49.136 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:15:26 +0000 (0:00:00.814)       0:12:49.951 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:15:27 +0000 (0:00:00.781)       0:12:50.733 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:15:27 +0000 (0:00:00.766)       0:12:51.499 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:15:28 +0000 (0:00:00.783)       0:12:52.283 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:15:29 +0000 (0:00:00.813)       0:12:53.097 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:15:30 +0000 (0:00:00.804)       0:12:53.902 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:15:31 +0000 (0:00:00.800)       0:12:54.702 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:15:31 +0000 (0:00:00.854)       0:12:55.556 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:15:32 +0000 (0:00:00.833)       0:12:56.390 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:15:33 +0000 (0:00:00.875)       0:12:57.266 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:15:34 +0000 (0:00:00.770)       0:12:58.037 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:15:35 +0000 (0:00:00.773)       0:12:58.811 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:15:35 +0000 (0:00:00.771)       0:12:59.582 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:15:36 +0000 (0:00:00.809)       0:13:00.392 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:15:37 +0000 (0:00:00.774)       0:13:01.167 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:15:38 +0000 (0:00:00.824)       0:13:01.992 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:15:39 +0000 (0:00:01.110)       0:13:03.102 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:15:40 +0000 (0:00:01.063)       0:13:04.165 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:15:41 +0000 (0:00:01.073)       0:13:05.239 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:15:42 +0000 (0:00:00.768)       0:13:06.008 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:15:43 +0000 (0:00:00.888)       0:13:06.896 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:15:44 +0000 (0:00:01.395)       0:13:08.292 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nThursday 26 March 2026  05:15:45 +0000 (0:00:00.865)       0:13:09.157 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nThursday 26 March 2026  05:15:46 +0000 (0:00:01.241)       0:13:10.398 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nThursday 26 March 2026  05:15:49 +0000 (0:00:03.163)       0:13:13.562 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nThursday 26 March 2026  05:15:51 +0000 (0:00:01.211)       0:13:14.774 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nThursday 26 March 2026  05:15:52 +0000 (0:00:01.156)       0:13:15.930 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nThursday 26 March 2026  05:15:53 +0000 (0:00:01.142)       0:13:17.073 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nThursday 26 March 2026  05:15:55 +0000 (0:00:02.017)       0:13:19.091 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nThursday 26 March 2026  05:15:56 +0000 (0:00:01.554)       0:13:20.646 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nThursday 26 March 2026  05:15:58 +0000 (0:00:01.544)       0:13:22.190 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nThursday 26 March 2026  05:16:00 +0000 (0:00:01.473)       0:13:23.663 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nThursday 26 March 2026  05:16:01 +0000 (0:00:01.616)       0:13:25.281 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nThursday 26 March 2026  05:16:03 +0000 (0:00:01.538)       0:13:26.819 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nThursday 26 March 2026  05:16:07 +0000 (0:00:04.191)       0:13:31.011 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nThursday 26 March 2026  05:16:09 +0000 (0:00:02.026)       0:13:33.038 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nThursday 26 March 2026  05:16:10 +0000 (0:00:01.175)       0:13:34.213 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nThursday 26 March 2026  05:16:11 +0000 (0:00:01.129)       0:13:35.343 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nThursday 26 March 2026  05:16:13 +0000 (0:00:01.741)       0:13:37.085 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nThursday 26 March 2026  05:16:15 +0000 (0:00:01.576)       0:13:38.661 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nThursday 26 March 2026  05:16:15 +0000 (0:00:00.808)       0:13:39.470 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:16:16 +0000 (0:00:01.167)       0:13:40.638 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nThursday 26 March 2026  05:16:18 +0000 (0:00:01.170)       0:13:41.809 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:16:19 +0000 (0:00:01.165)       0:13:42.974 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nThursday 26 March 2026  05:16:20 +0000 (0:00:01.140)       0:13:44.114 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nThursday 26 March 2026  05:16:23 +0000 (0:00:02.643)       0:13:46.758 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nThursday 26 March 2026  05:16:25 +0000 (0:00:01.972)       0:13:48.730 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nThursday 26 March 2026  05:16:27 +0000 (0:00:02.330)       0:13:51.061 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nThursday 26 March 2026  05:16:30 +0000 (0:00:03.186)       0:13:54.247 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nThursday 26 March 2026  05:16:31 +0000 (0:00:01.118)       0:13:55.366 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nThursday 26 March 2026  05:16:54 +0000 (0:00:22.816)       0:14:18.182 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nThursday 26 March 2026  05:16:57 +0000 (0:00:02.677)       0:14:20.859 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nThursday 26 March 2026  05:16:58 +0000 (0:00:00.807)       0:14:21.667 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m\n\nTASK [Start ceph mgr] **********************************************************\nThursday 26 March 2026  05:17:07 +0000 (0:00:09.327)       0:14:30.995 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:17:09 +0000 (0:00:02.117)       0:14:33.112 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:17:11 +0000 (0:00:01.877)       0:14:34.989 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Non container | waiting for the monitor to join the quorum...] ***********\nThursday 26 March 2026  05:17:12 +0000 (0:00:01.109)       0:14:36.099 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Container | waiting for the containerized monitor to join the quorum...] ***\nThursday 26 March 2026  05:17:13 +0000 (0:00:00.761)       0:14:36.861 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:17:15 +0000 (0:00:02.347)       0:14:39.208 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:17:16 +0000 (0:00:00.800)       0:14:40.008 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:17:17 +0000 (0:00:00.794)       0:14:40.803 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:17:17 +0000 (0:00:00.773)       0:14:41.577 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:17:18 +0000 (0:00:00.798)       0:14:42.375 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:17:19 +0000 (0:00:00.784)       0:14:43.160 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:17:20 +0000 (0:00:00.800)       0:14:43.960 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Upgrade ceph mon cluster] ************************************************\n\nTASK [Remove ceph aliases] *****************************************************\nThursday 26 March 2026  05:17:21 +0000 (0:00:00.955)       0:14:44.915 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set mon_host_count] ******************************************************\nThursday 26 March 2026  05:17:22 +0000 (0:00:01.157)       0:14:46.072 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Fail when less than three monitors] **************************************\nThursday 26 March 2026  05:17:23 +0000 (0:00:00.794)       0:14:46.867 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Select a running monitor] ************************************************\nThursday 26 March 2026  05:17:23 +0000 (0:00:00.779)       0:14:47.647 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:17:24 +0000 (0:00:00.820)       0:14:48.468 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:17:25 +0000 (0:00:01.154)       0:14:49.622 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:17:27 +0000 (0:00:01.440)       0:14:51.063 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:17:28 +0000 (0:00:01.110)       0:14:52.173 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:17:29 +0000 (0:00:01.443)       0:14:53.617 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:17:31 +0000 (0:00:01.175)       0:14:54.793 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:17:32 +0000 (0:00:01.169)       0:14:55.963 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:17:33 +0000 (0:00:01.280)       0:14:57.244 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:17:34 +0000 (0:00:01.231)       0:14:58.476 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:17:35 +0000 (0:00:01.169)       0:14:59.645 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:17:38 +0000 (0:00:02.218)       0:15:01.863 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:17:39 +0000 (0:00:01.300)       0:15:03.164 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:17:42 +0000 (0:00:03.182)       0:15:06.346 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:17:44 +0000 (0:00:01.424)       0:15:07.771 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:17:45 +0000 (0:00:01.660)       0:15:09.431 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:17:46 +0000 (0:00:01.154)       0:15:10.586 ******** \n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:17:40.373487', 'end': '2026-03-26 05:17:40.412891', 'delta': '0:00:00.039404', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:17:40.900551', 'end': '2026-03-26 05:17:40.944725', 'delta': '0:00:00.044174', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:17:41.454921', 'end': '2026-03-26 05:17:41.516677', 'delta': '0:00:00.061756', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:17:48 +0000 (0:00:01.214)       0:15:11.801 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:17:49 +0000 (0:00:01.310)       0:15:13.111 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:17:50 +0000 (0:00:01.285)       0:15:14.397 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:17:51 +0000 (0:00:01.202)       0:15:15.600 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:17:53 +0000 (0:00:02.005)       0:15:17.605 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:17:55 +0000 (0:00:01.157)       0:15:18.763 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:17:56 +0000 (0:00:01.150)       0:15:19.914 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:17:57 +0000 (0:00:01.207)       0:15:21.122 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:17:58 +0000 (0:00:01.132)       0:15:22.254 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:17:59 +0000 (0:00:01.178)       0:15:23.433 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:18:00 +0000 (0:00:01.137)       0:15:24.571 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:18:02 +0000 (0:00:01.158)       0:15:25.729 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:18:03 +0000 (0:00:01.142)       0:15:26.872 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:18:04 +0000 (0:00:01.146)       0:15:28.019 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:18:05 +0000 (0:00:01.140)       0:15:29.160 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:18:06 +0000 (0:00:01.292)       0:15:30.452 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:18:08 +0000 (0:00:01.222)       0:15:31.675 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:18:09 +0000 (0:00:01.485)       0:15:33.164 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:18:10 +0000 (0:00:01.164)       0:15:34.328 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:18:12 +0000 (0:00:01.508)       0:15:35.837 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:18:13 +0000 (0:00:01.109)       0:15:36.947 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:18:14 +0000 (0:00:01.278)       0:15:38.225 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:18:15 +0000 (0:00:01.176)       0:15:39.402 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:18:17 +0000 (0:00:01.686)       0:15:41.088 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:18:18 +0000 (0:00:01.213)       0:15:42.302 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:18:19 +0000 (0:00:01.119)       0:15:43.421 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:18:21 +0000 (0:00:01.904)       0:15:45.326 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get ceph cluster status] *************************************************\nThursday 26 March 2026  05:18:23 +0000 (0:00:02.305)       0:15:47.632 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Display ceph health detail] **********************************************\nThursday 26 March 2026  05:18:24 +0000 (0:00:00.919)       0:15:48.551 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Fail if cluster isn't in an acceptable state] ****************************\nThursday 26 March 2026  05:18:25 +0000 (0:00:00.932)       0:15:49.484 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Get the ceph quorum status] **********************************************\nThursday 26 March 2026  05:18:26 +0000 (0:00:00.789)       0:15:50.274 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Fail if the cluster quorum isn't in an acceptable state] *****************\nThursday 26 March 2026  05:18:27 +0000 (0:00:00.916)       0:15:51.191 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************\nThursday 26 March 2026  05:18:28 +0000 (0:00:00.769)       0:15:51.961 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Create potentially missing keys (rbd and rbd-mirror)] ********************\nThursday 26 March 2026  05:18:29 +0000 (0:00:01.408)       0:15:53.369 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-1']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-2']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-1']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-2']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Stop ceph mon] ***********************************************************\nThursday 26 March 2026  05:18:31 +0000 (0:00:01.623)       0:15:54.993 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [Mask the mgr service] ****************************************************\nThursday 26 March 2026  05:18:34 +0000 (0:00:03.431)       0:15:58.424 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:18:37 +0000 (0:00:02.252)       0:16:00.677 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:18:38 +0000 (0:00:01.081)       0:16:01.758 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:18:39 +0000 (0:00:01.146)       0:16:02.904 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:18:40 +0000 (0:00:01.594)       0:16:04.499 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:18:41 +0000 (0:00:01.137)       0:16:05.636 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:18:43 +0000 (0:00:01.137)       0:16:06.774 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:18:44 +0000 (0:00:01.107)       0:16:07.881 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:18:45 +0000 (0:00:01.586)       0:16:09.468 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:18:46 +0000 (0:00:01.119)       0:16:10.588 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:18:48 +0000 (0:00:01.187)       0:16:11.776 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:18:49 +0000 (0:00:01.609)       0:16:13.386 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:18:51 +0000 (0:00:01.563)       0:16:14.950 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:18:52 +0000 (0:00:00.777)       0:16:15.728 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:18:52 +0000 (0:00:00.767)       0:16:16.496 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:18:53 +0000 (0:00:00.800)       0:16:17.296 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:18:54 +0000 (0:00:00.802)       0:16:18.098 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:18:55 +0000 (0:00:00.789)       0:16:18.888 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:18:56 +0000 (0:00:00.775)       0:16:19.664 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:18:56 +0000 (0:00:00.779)       0:16:20.443 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:18:57 +0000 (0:00:00.810)       0:16:21.254 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:18:58 +0000 (0:00:00.800)       0:16:22.054 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:18:59 +0000 (0:00:00.792)       0:16:22.846 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:19:00 +0000 (0:00:00.822)       0:16:23.669 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:19:00 +0000 (0:00:00.769)       0:16:24.439 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:19:01 +0000 (0:00:00.769)       0:16:25.208 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:19:02 +0000 (0:00:00.799)       0:16:26.008 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:19:03 +0000 (0:00:00.755)       0:16:26.763 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:19:03 +0000 (0:00:00.746)       0:16:27.510 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:19:04 +0000 (0:00:00.769)       0:16:28.279 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:19:05 +0000 (0:00:00.778)       0:16:29.058 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:19:06 +0000 (0:00:00.765)       0:16:29.824 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:19:06 +0000 (0:00:00.758)       0:16:30.583 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:19:07 +0000 (0:00:00.757)       0:16:31.341 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:19:08 +0000 (0:00:00.767)       0:16:32.109 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:19:10 +0000 (0:00:01.645)       0:16:33.755 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:19:12 +0000 (0:00:02.045)       0:16:35.800 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:19:13 +0000 (0:00:01.119)       0:16:36.920 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:19:14 +0000 (0:00:01.130)       0:16:38.051 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:19:15 +0000 (0:00:01.163)       0:16:39.215 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:19:17 +0000 (0:00:01.897)       0:16:41.112 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:19:18 +0000 (0:00:01.455)       0:16:42.568 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:19:20 +0000 (0:00:01.125)       0:16:43.694 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:19:20 +0000 (0:00:00.803)       0:16:44.497 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:19:21 +0000 (0:00:00.774)       0:16:45.272 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:19:22 +0000 (0:00:01.137)       0:16:46.410 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:19:24 +0000 (0:00:01.892)       0:16:48.302 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:19:25 +0000 (0:00:01.145)       0:16:49.448 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:19:26 +0000 (0:00:01.136)       0:16:50.584 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:19:28 +0000 (0:00:01.197)       0:16:51.782 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:19:29 +0000 (0:00:01.157)       0:16:52.939 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:19:30 +0000 (0:00:01.179)       0:16:54.118 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:19:31 +0000 (0:00:00.818)       0:16:54.937 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:19:33 +0000 (0:00:02.279)       0:16:57.216 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:19:34 +0000 (0:00:00.763)       0:16:57.980 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:19:35 +0000 (0:00:01.138)       0:16:59.118 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:19:36 +0000 (0:00:01.132)       0:17:00.251 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:19:37 +0000 (0:00:01.140)       0:17:01.391 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:19:38 +0000 (0:00:01.184)       0:17:02.576 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:19:40 +0000 (0:00:01.154)       0:17:03.730 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:19:41 +0000 (0:00:01.219)       0:17:04.950 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:19:42 +0000 (0:00:01.154)       0:17:06.105 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:19:43 +0000 (0:00:01.176)       0:17:07.281 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:19:44 +0000 (0:00:01.206)       0:17:08.487 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:19:45 +0000 (0:00:00.849)       0:17:09.337 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:19:46 +0000 (0:00:01.103)       0:17:10.441 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:19:53 +0000 (0:00:06.354)       0:17:16.795 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:19:53 +0000 (0:00:00.763)       0:17:17.558 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:19:54 +0000 (0:00:00.774)       0:17:18.333 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:19:55 +0000 (0:00:00.800)       0:17:19.134 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:19:56 +0000 (0:00:00.817)       0:17:19.952 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:19:57 +0000 (0:00:00.834)       0:17:20.786 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:19:58 +0000 (0:00:00.894)       0:17:21.680 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:19:58 +0000 (0:00:00.847)       0:17:22.528 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:19:59 +0000 (0:00:00.773)       0:17:23.302 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:20:00 +0000 (0:00:00.800)       0:17:24.102 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:20:01 +0000 (0:00:00.767)       0:17:24.870 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:20:01 +0000 (0:00:00.777)       0:17:25.648 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:20:02 +0000 (0:00:00.786)       0:17:26.434 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:20:03 +0000 (0:00:00.894)       0:17:27.328 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:20:04 +0000 (0:00:00.770)       0:17:28.099 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:20:05 +0000 (0:00:00.882)       0:17:28.981 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:20:06 +0000 (0:00:00.758)       0:17:29.740 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:20:06 +0000 (0:00:00.784)       0:17:30.525 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:20:07 +0000 (0:00:00.743)       0:17:31.269 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:20:08 +0000 (0:00:00.780)       0:17:32.049 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:20:09 +0000 (0:00:00.762)       0:17:32.812 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:20:09 +0000 (0:00:00.748)       0:17:33.560 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:20:10 +0000 (0:00:01.070)       0:17:34.631 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:20:12 +0000 (0:00:01.039)       0:17:35.670 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:20:13 +0000 (0:00:01.016)       0:17:36.687 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:20:13 +0000 (0:00:00.772)       0:17:37.459 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:20:14 +0000 (0:00:00.929)       0:17:38.389 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:20:16 +0000 (0:00:01.440)       0:17:39.830 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nThursday 26 March 2026  05:20:17 +0000 (0:00:00.829)       0:17:40.660 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nThursday 26 March 2026  05:20:18 +0000 (0:00:01.164)       0:17:41.824 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nThursday 26 March 2026  05:20:21 +0000 (0:00:03.209)       0:17:45.033 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nThursday 26 March 2026  05:20:22 +0000 (0:00:01.188)       0:17:46.222 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nThursday 26 March 2026  05:20:23 +0000 (0:00:01.153)       0:17:47.375 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nThursday 26 March 2026  05:20:24 +0000 (0:00:01.164)       0:17:48.540 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nThursday 26 March 2026  05:20:26 +0000 (0:00:02.091)       0:17:50.631 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nThursday 26 March 2026  05:20:28 +0000 (0:00:01.605)       0:17:52.237 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nThursday 26 March 2026  05:20:30 +0000 (0:00:01.522)       0:17:53.759 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nThursday 26 March 2026  05:20:31 +0000 (0:00:01.533)       0:17:55.293 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nThursday 26 March 2026  05:20:33 +0000 (0:00:01.584)       0:17:56.877 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nThursday 26 March 2026  05:20:34 +0000 (0:00:01.548)       0:17:58.425 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nThursday 26 March 2026  05:20:38 +0000 (0:00:04.195)       0:18:02.621 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nThursday 26 March 2026  05:20:41 +0000 (0:00:02.120)       0:18:04.742 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nThursday 26 March 2026  05:20:42 +0000 (0:00:01.146)       0:18:05.888 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nThursday 26 March 2026  05:20:43 +0000 (0:00:01.151)       0:18:07.039 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nThursday 26 March 2026  05:20:45 +0000 (0:00:01.799)       0:18:08.838 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nThursday 26 March 2026  05:20:46 +0000 (0:00:01.465)       0:18:10.304 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nThursday 26 March 2026  05:20:47 +0000 (0:00:00.760)       0:18:11.065 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:20:48 +0000 (0:00:01.123)       0:18:12.189 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nThursday 26 March 2026  05:20:49 +0000 (0:00:01.118)       0:18:13.308 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:20:50 +0000 (0:00:01.176)       0:18:14.484 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nThursday 26 March 2026  05:20:51 +0000 (0:00:01.095)       0:18:15.579 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nThursday 26 March 2026  05:20:54 +0000 (0:00:02.713)       0:18:18.292 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nThursday 26 March 2026  05:20:56 +0000 (0:00:02.065)       0:18:20.358 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nThursday 26 March 2026  05:20:59 +0000 (0:00:02.468)       0:18:22.827 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nThursday 26 March 2026  05:21:03 +0000 (0:00:04.019)       0:18:26.846 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nThursday 26 March 2026  05:21:04 +0000 (0:00:01.151)       0:18:27.998 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nThursday 26 March 2026  05:21:27 +0000 (0:00:23.007)       0:18:51.006 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nThursday 26 March 2026  05:21:30 +0000 (0:00:02.689)       0:18:53.695 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nThursday 26 March 2026  05:21:30 +0000 (0:00:00.801)       0:18:54.496 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m\n\nTASK [Start ceph mgr] **********************************************************\nThursday 26 March 2026  05:21:40 +0000 (0:00:09.632)       0:19:04.129 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:21:42 +0000 (0:00:02.183)       0:19:06.312 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:21:44 +0000 (0:00:01.868)       0:19:08.181 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Non container | waiting for the monitor to join the quorum...] ***********\nThursday 26 March 2026  05:21:45 +0000 (0:00:01.420)       0:19:09.602 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Container | waiting for the containerized monitor to join the quorum...] ***\nThursday 26 March 2026  05:21:46 +0000 (0:00:00.762)       0:19:10.364 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:21:48 +0000 (0:00:01.963)       0:19:12.328 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:21:49 +0000 (0:00:00.825)       0:19:13.154 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:21:50 +0000 (0:00:00.762)       0:19:13.916 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:21:51 +0000 (0:00:00.773)       0:19:14.690 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:21:51 +0000 (0:00:00.754)       0:19:15.444 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:21:52 +0000 (0:00:00.764)       0:19:16.209 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:21:53 +0000 (0:00:00.845)       0:19:17.055 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reset mon_host] **********************************************************\n\nTASK [Reset mon_host fact] *****************************************************\nThursday 26 March 2026  05:21:55 +0000 (0:00:01.781)       0:19:18.836 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:21:56 +0000 (0:00:01.460)       0:19:20.297 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:21:57 +0000 (0:00:01.175)       0:19:21.473 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:21:58 +0000 (0:00:01.170)       0:19:22.644 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:22:00 +0000 (0:00:01.141)       0:19:23.785 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:22:01 +0000 (0:00:01.174)       0:19:24.959 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:22:02 +0000 (0:00:01.123)       0:19:26.083 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:22:03 +0000 (0:00:01.125)       0:19:27.208 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:22:04 +0000 (0:00:01.125)       0:19:28.334 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:22:05 +0000 (0:00:01.134)       0:19:29.469 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:22:06 +0000 (0:00:01.108)       0:19:30.578 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:22:08 +0000 (0:00:01.120)       0:19:31.698 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:22:09 +0000 (0:00:01.107)       0:19:32.806 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:22:10 +0000 (0:00:01.133)       0:19:33.939 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:22:11 +0000 (0:00:01.194)       0:19:35.133 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:22:12 +0000 (0:00:01.154)       0:19:36.288 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:22:13 +0000 (0:00:01.114)       0:19:37.402 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:22:14 +0000 (0:00:01.123)       0:19:38.526 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:22:15 +0000 (0:00:01.121)       0:19:39.648 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:22:17 +0000 (0:00:01.103)       0:19:40.751 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:22:18 +0000 (0:00:01.113)       0:19:41.865 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:22:19 +0000 (0:00:01.172)       0:19:43.038 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:22:20 +0000 (0:00:01.113)       0:19:44.151 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:22:21 +0000 (0:00:01.187)       0:19:45.339 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:22:22 +0000 (0:00:01.154)       0:19:46.494 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:22:23 +0000 (0:00:01.131)       0:19:47.625 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:22:25 +0000 (0:00:01.147)       0:19:48.773 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:22:26 +0000 (0:00:01.146)       0:19:49.919 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:22:27 +0000 (0:00:01.144)       0:19:51.064 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:22:28 +0000 (0:00:01.208)       0:19:52.273 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:22:29 +0000 (0:00:01.150)       0:19:53.424 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:22:30 +0000 (0:00:01.177)       0:19:54.602 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:22:32 +0000 (0:00:01.125)       0:19:55.728 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:22:33 +0000 (0:00:01.243)       0:19:56.972 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:22:34 +0000 (0:00:01.092)       0:19:58.064 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:22:35 +0000 (0:00:01.138)       0:19:59.203 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:22:36 +0000 (0:00:01.115)       0:20:00.318 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:22:37 +0000 (0:00:01.113)       0:20:01.432 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:22:38 +0000 (0:00:01.132)       0:20:02.564 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:22:40 +0000 (0:00:01.128)       0:20:03.692 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:22:41 +0000 (0:00:01.147)       0:20:04.840 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:22:42 +0000 (0:00:01.204)       0:20:06.045 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:22:43 +0000 (0:00:01.141)       0:20:07.187 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:22:44 +0000 (0:00:01.124)       0:20:08.312 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:22:45 +0000 (0:00:01.109)       0:20:09.422 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:22:46 +0000 (0:00:01.096)       0:20:10.518 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:22:48 +0000 (0:00:01.165)       0:20:11.684 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:22:49 +0000 (0:00:01.245)       0:20:12.929 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:22:50 +0000 (0:00:01.181)       0:20:14.110 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:22:51 +0000 (0:00:01.283)       0:20:15.393 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:22:52 +0000 (0:00:01.180)       0:20:16.574 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:22:54 +0000 (0:00:01.124)       0:20:17.698 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:22:55 +0000 (0:00:01.155)       0:20:18.853 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:22:56 +0000 (0:00:01.160)       0:20:20.014 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:22:57 +0000 (0:00:01.128)       0:20:21.143 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:22:58 +0000 (0:00:01.155)       0:20:22.298 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:23:00 +0000 (0:00:01.393)       0:20:23.692 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:23:01 +0000 (0:00:01.760)       0:20:25.452 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:23:03 +0000 (0:00:01.710)       0:20:27.162 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:23:04 +0000 (0:00:01.196)       0:20:28.359 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:23:05 +0000 (0:00:01.236)       0:20:29.596 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:23:07 +0000 (0:00:01.150)       0:20:30.746 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:23:08 +0000 (0:00:01.401)       0:20:32.148 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:23:09 +0000 (0:00:01.170)       0:20:33.318 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:23:10 +0000 (0:00:01.141)       0:20:34.460 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:23:11 +0000 (0:00:01.169)       0:20:35.630 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:23:12 +0000 (0:00:00.967)       0:20:36.598 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:23:13 +0000 (0:00:00.781)       0:20:37.380 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:23:14 +0000 (0:00:00.869)       0:20:38.249 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:23:15 +0000 (0:00:00.820)       0:20:39.070 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:23:16 +0000 (0:00:00.771)       0:20:39.842 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:23:16 +0000 (0:00:00.811)       0:20:40.653 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:23:17 +0000 (0:00:00.792)       0:20:41.446 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:23:18 +0000 (0:00:00.772)       0:20:42.218 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:23:19 +0000 (0:00:00.782)       0:20:43.001 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:23:20 +0000 (0:00:00.798)       0:20:43.799 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:23:20 +0000 (0:00:00.797)       0:20:44.597 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:23:21 +0000 (0:00:00.806)       0:20:45.403 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:23:22 +0000 (0:00:00.787)       0:20:46.191 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:23:23 +0000 (0:00:00.808)       0:20:47.000 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:23:24 +0000 (0:00:00.810)       0:20:47.810 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:23:24 +0000 (0:00:00.780)       0:20:48.591 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:23:25 +0000 (0:00:00.798)       0:20:49.389 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:23:26 +0000 (0:00:00.784)       0:20:50.174 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:23:27 +0000 (0:00:00.815)       0:20:50.989 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:23:28 +0000 (0:00:00.775)       0:20:51.765 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:23:28 +0000 (0:00:00.785)       0:20:52.550 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:23:29 +0000 (0:00:00.793)       0:20:53.343 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:23:30 +0000 (0:00:00.771)       0:20:54.115 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:23:31 +0000 (0:00:00.776)       0:20:54.892 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:23:32 +0000 (0:00:00.767)       0:20:55.659 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:23:32 +0000 (0:00:00.799)       0:20:56.459 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:23:33 +0000 (0:00:00.835)       0:20:57.295 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:23:34 +0000 (0:00:00.813)       0:20:58.108 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:23:35 +0000 (0:00:00.786)       0:20:58.895 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:23:36 +0000 (0:00:00.797)       0:20:59.693 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:23:36 +0000 (0:00:00.758)       0:21:00.451 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:23:37 +0000 (0:00:00.782)       0:21:01.234 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:23:38 +0000 (0:00:00.824)       0:21:02.058 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:23:39 +0000 (0:00:00.784)       0:21:02.843 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:23:39 +0000 (0:00:00.767)       0:21:03.610 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:23:40 +0000 (0:00:00.867)       0:21:04.478 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:23:41 +0000 (0:00:00.785)       0:21:05.263 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:23:42 +0000 (0:00:00.801)       0:21:06.065 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:23:43 +0000 (0:00:00.780)       0:21:06.845 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:23:43 +0000 (0:00:00.780)       0:21:07.625 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:23:44 +0000 (0:00:00.836)       0:21:08.462 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:23:45 +0000 (0:00:00.786)       0:21:09.249 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:23:46 +0000 (0:00:00.763)       0:21:10.012 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:23:47 +0000 (0:00:00.797)       0:21:10.810 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:23:47 +0000 (0:00:00.761)       0:21:11.572 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:23:48 +0000 (0:00:00.781)       0:21:12.353 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:23:49 +0000 (0:00:00.898)       0:21:13.252 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:23:50 +0000 (0:00:00.781)       0:21:14.033 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:23:51 +0000 (0:00:00.885)       0:21:14.919 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:23:52 +0000 (0:00:00.816)       0:21:15.735 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:23:52 +0000 (0:00:00.775)       0:21:16.511 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:23:53 +0000 (0:00:00.787)       0:21:17.298 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:23:54 +0000 (0:00:00.806)       0:21:18.104 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:23:55 +0000 (0:00:00.815)       0:21:18.920 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:23:56 +0000 (0:00:00.816)       0:21:19.737 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:23:57 +0000 (0:00:01.025)       0:21:20.762 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:23:58 +0000 (0:00:01.084)       0:21:21.846 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:23:59 +0000 (0:00:01.080)       0:21:22.927 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:24:00 +0000 (0:00:00.766)       0:21:23.693 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:24:00 +0000 (0:00:00.911)       0:21:24.605 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:24:01 +0000 (0:00:00.804)       0:21:25.410 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:24:03 +0000 (0:00:01.438)       0:21:26.848 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:24:03 +0000 (0:00:00.788)       0:21:27.637 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:24:04 +0000 (0:00:00.815)       0:21:28.452 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:24:05 +0000 (0:00:00.790)       0:21:29.243 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:24:06 +0000 (0:00:01.318)       0:21:30.561 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:24:07 +0000 (0:00:00.785)       0:21:31.347 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:24:08 +0000 (0:00:00.769)       0:21:32.116 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:24:09 +0000 (0:00:00.788)       0:21:32.905 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:24:10 +0000 (0:00:00.760)       0:21:33.665 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:24:10 +0000 (0:00:00.800)       0:21:34.466 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:24:11 +0000 (0:00:00.767)       0:21:35.233 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:24:12 +0000 (0:00:00.764)       0:21:35.998 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:24:13 +0000 (0:00:00.774)       0:21:36.772 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:24:13 +0000 (0:00:00.810)       0:21:37.583 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:24:14 +0000 (0:00:00.778)       0:21:38.362 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:24:15 +0000 (0:00:00.781)       0:21:39.144 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:24:16 +0000 (0:00:00.772)       0:21:39.916 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:24:17 +0000 (0:00:00.838)       0:21:40.755 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:24:17 +0000 (0:00:00.785)       0:21:41.540 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:24:18 +0000 (0:00:00.772)       0:21:42.313 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:24:19 +0000 (0:00:00.789)       0:21:43.102 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:24:20 +0000 (0:00:00.770)       0:21:43.873 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:24:20 +0000 (0:00:00.758)       0:21:44.632 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:24:21 +0000 (0:00:00.793)       0:21:45.425 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:24:22 +0000 (0:00:00.833)       0:21:46.258 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:24:23 +0000 (0:00:00.781)       0:21:47.040 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:24:24 +0000 (0:00:00.810)       0:21:47.851 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:24:24 +0000 (0:00:00.778)       0:21:48.630 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:24:25 +0000 (0:00:00.769)       0:21:49.399 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:24:26 +0000 (0:00:00.815)       0:21:50.215 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:24:27 +0000 (0:00:00.807)       0:21:51.023 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:24:28 +0000 (0:00:00.766)       0:21:51.789 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:24:28 +0000 (0:00:00.771)       0:21:52.560 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:24:29 +0000 (0:00:00.755)       0:21:53.316 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:24:30 +0000 (0:00:00.756)       0:21:54.072 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:24:31 +0000 (0:00:00.767)       0:21:54.840 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:24:32 +0000 (0:00:00.897)       0:21:55.738 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:24:32 +0000 (0:00:00.824)       0:21:56.563 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:24:33 +0000 (0:00:00.821)       0:21:57.385 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:24:34 +0000 (0:00:00.776)       0:21:58.162 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:24:35 +0000 (0:00:00.772)       0:21:58.935 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:24:36 +0000 (0:00:00.775)       0:21:59.710 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:24:36 +0000 (0:00:00.794)       0:22:00.505 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:24:37 +0000 (0:00:00.774)       0:22:01.279 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:24:38 +0000 (0:00:00.774)       0:22:02.054 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:24:39 +0000 (0:00:00.806)       0:22:02.860 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:24:39 +0000 (0:00:00.757)       0:22:03.617 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:24:40 +0000 (0:00:00.791)       0:22:04.409 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:24:41 +0000 (0:00:00.815)       0:22:05.225 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:24:42 +0000 (0:00:00.811)       0:22:06.036 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:24:43 +0000 (0:00:00.899)       0:22:06.936 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:24:44 +0000 (0:00:00.782)       0:22:07.719 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:24:44 +0000 (0:00:00.858)       0:22:08.577 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:24:45 +0000 (0:00:00.774)       0:22:09.352 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:24:46 +0000 (0:00:00.814)       0:22:10.166 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:24:47 +0000 (0:00:00.837)       0:22:11.004 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:24:48 +0000 (0:00:00.763)       0:22:11.767 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:24:48 +0000 (0:00:00.787)       0:22:12.555 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:24:49 +0000 (0:00:00.774)       0:22:13.330 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:24:50 +0000 (0:00:01.089)       0:22:14.419 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:24:52 +0000 (0:00:01.469)       0:22:15.889 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:24:53 +0000 (0:00:01.376)       0:22:17.266 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:24:54 +0000 (0:00:00.795)       0:22:18.061 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:24:55 +0000 (0:00:00.901)       0:22:18.963 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:24:56 +0000 (0:00:00.795)       0:22:19.758 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:24:57 +0000 (0:00:01.055)       0:22:20.814 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:24:57 +0000 (0:00:00.778)       0:22:21.593 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:24:58 +0000 (0:00:00.766)       0:22:22.360 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:24:59 +0000 (0:00:00.809)       0:22:23.170 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes] **************************************************\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:25:00 +0000 (0:00:01.409)       0:22:24.579 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Mask ceph mgr systemd unit] **********************************************\nThursday 26 March 2026  05:25:13 +0000 (0:00:12.895)       0:22:37.474 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:25:16 +0000 (0:00:02.590)       0:22:40.065 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:25:17 +0000 (0:00:01.141)       0:22:41.206 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:25:19 +0000 (0:00:01.524)       0:22:42.731 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:25:20 +0000 (0:00:01.146)       0:22:43.877 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:25:21 +0000 (0:00:01.532)       0:22:45.410 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:25:22 +0000 (0:00:01.118)       0:22:46.528 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:25:24 +0000 (0:00:01.136)       0:22:47.665 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:25:25 +0000 (0:00:01.154)       0:22:48.820 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:25:26 +0000 (0:00:01.157)       0:22:49.977 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:25:27 +0000 (0:00:01.099)       0:22:51.077 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:25:29 +0000 (0:00:01.962)       0:22:53.040 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:25:30 +0000 (0:00:01.203)       0:22:54.244 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:25:33 +0000 (0:00:03.180)       0:22:57.424 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:25:35 +0000 (0:00:01.779)       0:22:59.203 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:25:37 +0000 (0:00:01.665)       0:23:00.869 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:25:38 +0000 (0:00:01.192)       0:23:02.061 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:25:31.452374', 'end': '2026-03-26 05:25:31.504874', 'delta': '0:00:00.052500', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:25:32.004082', 'end': '2026-03-26 05:25:32.055136', 'delta': '0:00:00.051054', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:25:32.553894', 'end': '2026-03-26 05:25:32.609169', 'delta': '0:00:00.055275', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:25:39 +0000 (0:00:01.241)       0:23:03.303 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:25:40 +0000 (0:00:01.309)       0:23:04.612 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:25:42 +0000 (0:00:01.286)       0:23:05.898 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:25:43 +0000 (0:00:01.138)       0:23:07.037 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:25:45 +0000 (0:00:02.009)       0:23:09.046 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:25:46 +0000 (0:00:01.180)       0:23:10.226 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:25:47 +0000 (0:00:01.129)       0:23:11.356 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:25:48 +0000 (0:00:01.190)       0:23:12.547 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:25:50 +0000 (0:00:01.181)       0:23:13.728 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:25:51 +0000 (0:00:01.150)       0:23:14.879 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:25:52 +0000 (0:00:01.189)       0:23:16.069 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:25:53 +0000 (0:00:01.160)       0:23:17.229 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:25:54 +0000 (0:00:01.140)       0:23:18.370 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:25:55 +0000 (0:00:01.182)       0:23:19.553 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:25:57 +0000 (0:00:01.126)       0:23:20.679 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:25:58 +0000 (0:00:01.346)       0:23:22.026 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:25:59 +0000 (0:00:01.217)       0:23:23.244 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:26:01 +0000 (0:00:01.553)       0:23:24.797 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:26:02 +0000 (0:00:01.183)       0:23:25.980 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:26:03 +0000 (0:00:01.524)       0:23:27.505 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:26:04 +0000 (0:00:01.131)       0:23:28.636 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:26:06 +0000 (0:00:01.244)       0:23:29.880 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:26:07 +0000 (0:00:01.138)       0:23:31.019 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:26:09 +0000 (0:00:02.118)       0:23:33.138 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:26:10 +0000 (0:00:01.158)       0:23:34.296 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:26:11 +0000 (0:00:01.167)       0:23:35.463 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:26:13 +0000 (0:00:01.837)       0:23:37.300 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:26:16 +0000 (0:00:02.588)       0:23:39.889 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:26:17 +0000 (0:00:01.119)       0:23:41.009 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:26:18 +0000 (0:00:01.097)       0:23:42.107 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:26:20 +0000 (0:00:01.558)       0:23:43.665 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:26:21 +0000 (0:00:01.112)       0:23:44.778 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:26:22 +0000 (0:00:01.139)       0:23:45.918 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:26:23 +0000 (0:00:01.106)       0:23:47.024 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:26:24 +0000 (0:00:01.584)       0:23:48.609 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:26:26 +0000 (0:00:01.162)       0:23:49.772 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:26:27 +0000 (0:00:01.154)       0:23:50.927 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:26:28 +0000 (0:00:01.595)       0:23:52.522 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:26:30 +0000 (0:00:01.566)       0:23:54.089 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:26:31 +0000 (0:00:01.126)       0:23:55.216 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:26:32 +0000 (0:00:01.179)       0:23:56.396 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:26:33 +0000 (0:00:01.135)       0:23:57.531 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:26:34 +0000 (0:00:01.113)       0:23:58.644 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:26:36 +0000 (0:00:01.131)       0:23:59.776 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:26:37 +0000 (0:00:01.113)       0:24:00.890 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:26:38 +0000 (0:00:01.143)       0:24:02.033 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:26:39 +0000 (0:00:01.229)       0:24:03.263 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:26:40 +0000 (0:00:01.177)       0:24:04.441 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:26:41 +0000 (0:00:01.178)       0:24:05.619 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:26:43 +0000 (0:00:01.167)       0:24:06.788 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:26:44 +0000 (0:00:01.177)       0:24:07.965 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:26:45 +0000 (0:00:01.153)       0:24:09.119 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:26:46 +0000 (0:00:01.126)       0:24:10.245 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:26:47 +0000 (0:00:01.145)       0:24:11.390 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:26:48 +0000 (0:00:01.144)       0:24:12.535 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:26:50 +0000 (0:00:01.167)       0:24:13.702 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:26:51 +0000 (0:00:01.168)       0:24:14.871 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:26:52 +0000 (0:00:01.152)       0:24:16.024 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:26:53 +0000 (0:00:01.171)       0:24:17.195 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:26:54 +0000 (0:00:01.142)       0:24:18.337 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:26:55 +0000 (0:00:01.128)       0:24:19.466 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:26:57 +0000 (0:00:01.990)       0:24:21.457 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:27:00 +0000 (0:00:02.349)       0:24:23.807 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:27:01 +0000 (0:00:01.169)       0:24:24.976 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:27:02 +0000 (0:00:01.153)       0:24:26.129 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:27:03 +0000 (0:00:01.119)       0:24:27.249 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:27:05 +0000 (0:00:01.818)       0:24:29.067 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:27:06 +0000 (0:00:01.522)       0:24:30.590 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:27:08 +0000 (0:00:01.244)       0:24:31.834 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:27:09 +0000 (0:00:01.141)       0:24:32.976 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:27:10 +0000 (0:00:01.148)       0:24:34.125 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:27:11 +0000 (0:00:01.260)       0:24:35.385 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:27:13 +0000 (0:00:01.823)       0:24:37.208 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:27:14 +0000 (0:00:01.112)       0:24:38.321 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:27:15 +0000 (0:00:01.131)       0:24:39.452 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:27:16 +0000 (0:00:01.147)       0:24:40.599 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:27:18 +0000 (0:00:01.135)       0:24:41.735 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:27:19 +0000 (0:00:01.173)       0:24:42.909 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:27:20 +0000 (0:00:01.175)       0:24:44.084 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:27:23 +0000 (0:00:02.611)       0:24:46.696 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:27:24 +0000 (0:00:01.159)       0:24:47.855 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:27:25 +0000 (0:00:01.298)       0:24:49.154 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:27:26 +0000 (0:00:01.141)       0:24:50.295 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:27:27 +0000 (0:00:01.144)       0:24:51.440 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:27:28 +0000 (0:00:01.127)       0:24:52.568 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:27:30 +0000 (0:00:01.194)       0:24:53.763 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:27:31 +0000 (0:00:01.142)       0:24:54.905 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:27:32 +0000 (0:00:01.173)       0:24:56.079 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:27:33 +0000 (0:00:01.149)       0:24:57.228 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:27:34 +0000 (0:00:01.161)       0:24:58.390 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:27:35 +0000 (0:00:01.158)       0:24:59.548 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:27:37 +0000 (0:00:01.118)       0:25:00.667 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:27:44 +0000 (0:00:07.036)       0:25:07.703 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:27:45 +0000 (0:00:01.134)       0:25:08.837 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:27:46 +0000 (0:00:01.117)       0:25:09.955 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:27:47 +0000 (0:00:01.153)       0:25:11.109 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:27:48 +0000 (0:00:01.133)       0:25:12.243 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:27:49 +0000 (0:00:01.139)       0:25:13.382 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:27:50 +0000 (0:00:01.108)       0:25:14.490 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:27:51 +0000 (0:00:01.102)       0:25:15.593 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:27:53 +0000 (0:00:01.116)       0:25:16.710 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:27:54 +0000 (0:00:01.112)       0:25:17.823 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:27:55 +0000 (0:00:01.125)       0:25:18.949 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:27:56 +0000 (0:00:01.133)       0:25:20.082 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:27:57 +0000 (0:00:01.154)       0:25:21.237 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:27:58 +0000 (0:00:01.192)       0:25:22.429 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:27:59 +0000 (0:00:01.157)       0:25:23.586 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:28:01 +0000 (0:00:01.294)       0:25:24.880 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:28:02 +0000 (0:00:01.262)       0:25:26.143 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:28:03 +0000 (0:00:01.094)       0:25:27.238 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:28:04 +0000 (0:00:01.118)       0:25:28.357 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:28:05 +0000 (0:00:01.147)       0:25:29.504 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:28:06 +0000 (0:00:00.973)       0:25:30.477 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:28:07 +0000 (0:00:01.029)       0:25:31.507 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:28:09 +0000 (0:00:01.337)       0:25:32.844 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:28:10 +0000 (0:00:01.380)       0:25:34.225 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:28:12 +0000 (0:00:01.607)       0:25:35.833 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:28:13 +0000 (0:00:01.149)       0:25:36.982 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:28:14 +0000 (0:00:01.442)       0:25:38.424 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:28:16 +0000 (0:00:01.846)       0:25:40.271 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:28:18 +0000 (0:00:01.689)       0:25:41.960 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nThursday 26 March 2026  05:28:19 +0000 (0:00:01.473)       0:25:43.434 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nThursday 26 March 2026  05:28:21 +0000 (0:00:01.533)       0:25:44.967 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nThursday 26 March 2026  05:28:22 +0000 (0:00:01.134)       0:25:46.102 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nThursday 26 March 2026  05:28:30 +0000 (0:00:07.644)       0:25:53.746 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:28:31 +0000 (0:00:01.177)       0:25:54.923 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:28:34 +0000 (0:00:03.200)       0:25:58.124 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nThursday 26 March 2026  05:28:36 +0000 (0:00:02.001)       0:26:00.125 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nThursday 26 March 2026  05:28:38 +0000 (0:00:01.570)       0:26:01.696 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:28:39 +0000 (0:00:01.205)       0:26:02.902 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:28:40 +0000 (0:00:01.138)       0:26:04.040 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:28:41 +0000 (0:00:01.469)       0:26:05.510 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nThursday 26 March 2026  05:28:43 +0000 (0:00:01.168)       0:26:06.679 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:28:44 +0000 (0:00:01.120)       0:26:07.800 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:28:45 +0000 (0:00:01.482)       0:26:09.283 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nThursday 26 March 2026  05:28:47 +0000 (0:00:02.048)       0:26:11.331 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nThursday 26 March 2026  05:28:50 +0000 (0:00:02.472)       0:26:13.803 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nThursday 26 March 2026  05:28:52 +0000 (0:00:02.585)       0:26:16.389 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:28:56 +0000 (0:00:03.817)       0:26:20.207 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes] **************************************************\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:28:57 +0000 (0:00:01.285)       0:26:21.492 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Mask ceph mgr systemd unit] **********************************************\nThursday 26 March 2026  05:29:00 +0000 (0:00:02.406)       0:26:23.899 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:29:02 +0000 (0:00:02.020)       0:26:25.919 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:29:03 +0000 (0:00:01.115)       0:26:27.035 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:29:04 +0000 (0:00:01.507)       0:26:28.542 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:29:06 +0000 (0:00:01.141)       0:26:29.683 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:29:07 +0000 (0:00:01.498)       0:26:31.182 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:29:08 +0000 (0:00:01.193)       0:26:32.376 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:29:09 +0000 (0:00:01.131)       0:26:33.508 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:29:10 +0000 (0:00:01.121)       0:26:34.630 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:29:12 +0000 (0:00:01.212)       0:26:35.842 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:29:13 +0000 (0:00:01.130)       0:26:36.973 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:29:15 +0000 (0:00:01.718)       0:26:38.692 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:29:16 +0000 (0:00:01.295)       0:26:39.988 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:29:19 +0000 (0:00:02.878)       0:26:42.867 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:29:20 +0000 (0:00:01.387)       0:26:44.254 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:29:22 +0000 (0:00:01.674)       0:26:45.929 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:29:23 +0000 (0:00:01.170)       0:26:47.099 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:29:16.907211', 'end': '2026-03-26 05:29:16.951545', 'delta': '0:00:00.044334', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:29:17.432469', 'end': '2026-03-26 05:29:17.476218', 'delta': '0:00:00.043749', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:29:17.941866', 'end': '2026-03-26 05:29:17.979790', 'delta': '0:00:00.037924', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:29:24 +0000 (0:00:01.215)       0:26:48.315 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:29:25 +0000 (0:00:01.296)       0:26:49.611 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:29:27 +0000 (0:00:01.226)       0:26:50.838 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:29:28 +0000 (0:00:01.159)       0:26:51.997 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:29:30 +0000 (0:00:02.059)       0:26:54.057 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:29:31 +0000 (0:00:01.115)       0:26:55.173 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:29:32 +0000 (0:00:01.191)       0:26:56.364 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:29:33 +0000 (0:00:01.258)       0:26:57.622 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:29:35 +0000 (0:00:01.136)       0:26:58.759 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:29:36 +0000 (0:00:01.140)       0:26:59.900 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:29:37 +0000 (0:00:01.161)       0:27:01.062 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:29:38 +0000 (0:00:01.142)       0:27:02.205 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:29:39 +0000 (0:00:01.142)       0:27:03.347 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:29:40 +0000 (0:00:01.148)       0:27:04.496 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:29:41 +0000 (0:00:01.160)       0:27:05.657 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:29:43 +0000 (0:00:01.294)       0:27:06.951 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:29:44 +0000 (0:00:01.259)       0:27:08.210 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:29:46 +0000 (0:00:01.563)       0:27:09.774 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:29:47 +0000 (0:00:01.115)       0:27:10.889 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:29:48 +0000 (0:00:01.501)       0:27:12.390 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:29:49 +0000 (0:00:01.148)       0:27:13.539 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:29:51 +0000 (0:00:01.275)       0:27:14.815 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:29:52 +0000 (0:00:01.166)       0:27:15.982 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:29:53 +0000 (0:00:01.664)       0:27:17.646 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:29:55 +0000 (0:00:01.160)       0:27:18.807 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:29:56 +0000 (0:00:01.127)       0:27:19.935 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:29:58 +0000 (0:00:02.117)       0:27:22.052 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:30:00 +0000 (0:00:02.294)       0:27:24.347 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:30:01 +0000 (0:00:01.124)       0:27:25.471 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:30:03 +0000 (0:00:01.317)       0:27:26.789 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:30:04 +0000 (0:00:01.521)       0:27:28.310 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:30:05 +0000 (0:00:01.166)       0:27:29.477 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:30:06 +0000 (0:00:01.083)       0:27:30.561 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:30:07 +0000 (0:00:01.080)       0:27:31.641 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:30:09 +0000 (0:00:01.550)       0:27:33.192 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:30:10 +0000 (0:00:01.089)       0:27:34.281 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:30:11 +0000 (0:00:01.108)       0:27:35.390 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:30:13 +0000 (0:00:01.582)       0:27:36.972 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:30:14 +0000 (0:00:01.528)       0:27:38.500 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:30:15 +0000 (0:00:00.756)       0:27:39.257 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:30:16 +0000 (0:00:00.834)       0:27:40.091 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:30:17 +0000 (0:00:00.774)       0:27:40.865 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:30:17 +0000 (0:00:00.756)       0:27:41.622 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:30:18 +0000 (0:00:00.810)       0:27:42.432 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:30:19 +0000 (0:00:00.836)       0:27:43.269 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:30:20 +0000 (0:00:00.763)       0:27:44.032 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:30:21 +0000 (0:00:00.823)       0:27:44.856 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:30:21 +0000 (0:00:00.788)       0:27:45.645 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:30:22 +0000 (0:00:00.828)       0:27:46.474 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:30:23 +0000 (0:00:00.808)       0:27:47.282 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:30:24 +0000 (0:00:00.780)       0:27:48.063 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:30:25 +0000 (0:00:00.780)       0:27:48.843 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:30:25 +0000 (0:00:00.789)       0:27:49.633 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:30:26 +0000 (0:00:00.777)       0:27:50.410 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:30:27 +0000 (0:00:00.784)       0:27:51.195 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:30:28 +0000 (0:00:00.825)       0:27:52.020 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:30:29 +0000 (0:00:00.765)       0:27:52.786 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:30:29 +0000 (0:00:00.747)       0:27:53.533 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:30:30 +0000 (0:00:00.772)       0:27:54.306 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:30:31 +0000 (0:00:00.765)       0:27:55.072 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:30:32 +0000 (0:00:00.815)       0:27:55.887 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:30:33 +0000 (0:00:01.567)       0:27:57.455 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:30:35 +0000 (0:00:01.940)       0:27:59.395 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:30:36 +0000 (0:00:01.097)       0:28:00.493 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:30:38 +0000 (0:00:01.191)       0:28:01.685 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:30:39 +0000 (0:00:01.120)       0:28:02.806 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:30:40 +0000 (0:00:01.829)       0:28:04.636 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:30:42 +0000 (0:00:01.533)       0:28:06.169 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:30:43 +0000 (0:00:01.186)       0:28:07.356 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:30:44 +0000 (0:00:00.764)       0:28:08.120 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:30:45 +0000 (0:00:00.795)       0:28:08.916 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:30:46 +0000 (0:00:01.216)       0:28:10.133 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:30:48 +0000 (0:00:01.843)       0:28:11.977 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:30:49 +0000 (0:00:01.130)       0:28:13.107 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:30:50 +0000 (0:00:01.157)       0:28:14.265 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:30:51 +0000 (0:00:01.189)       0:28:15.455 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:30:52 +0000 (0:00:01.150)       0:28:16.606 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:30:54 +0000 (0:00:01.164)       0:28:17.770 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:30:54 +0000 (0:00:00.777)       0:28:18.548 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:30:57 +0000 (0:00:02.187)       0:28:20.735 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:30:57 +0000 (0:00:00.812)       0:28:21.548 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:30:59 +0000 (0:00:01.166)       0:28:22.715 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:31:00 +0000 (0:00:01.135)       0:28:23.850 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:31:01 +0000 (0:00:01.143)       0:28:24.994 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:31:02 +0000 (0:00:01.145)       0:28:26.140 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:31:03 +0000 (0:00:01.120)       0:28:27.261 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:31:04 +0000 (0:00:01.113)       0:28:28.374 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:31:05 +0000 (0:00:01.153)       0:28:29.528 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:31:06 +0000 (0:00:01.126)       0:28:30.654 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:31:08 +0000 (0:00:01.164)       0:28:31.819 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:31:09 +0000 (0:00:00.941)       0:28:32.761 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:31:10 +0000 (0:00:01.112)       0:28:33.874 ******** \n\u001b[0;32mok: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:31:16 +0000 (0:00:06.138)       0:28:40.013 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:31:17 +0000 (0:00:00.765)       0:28:40.779 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:31:17 +0000 (0:00:00.768)       0:28:41.547 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:31:18 +0000 (0:00:00.801)       0:28:42.348 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:31:19 +0000 (0:00:00.784)       0:28:43.133 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:31:20 +0000 (0:00:00.771)       0:28:43.905 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:31:21 +0000 (0:00:00.787)       0:28:44.693 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:31:21 +0000 (0:00:00.750)       0:28:45.443 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:31:22 +0000 (0:00:00.821)       0:28:46.265 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:31:23 +0000 (0:00:00.844)       0:28:47.110 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:31:24 +0000 (0:00:00.797)       0:28:47.907 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:31:25 +0000 (0:00:00.772)       0:28:48.680 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:31:25 +0000 (0:00:00.766)       0:28:49.447 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:31:26 +0000 (0:00:00.900)       0:28:50.347 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:31:27 +0000 (0:00:00.764)       0:28:51.112 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:31:28 +0000 (0:00:00.869)       0:28:51.981 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:31:29 +0000 (0:00:00.761)       0:28:52.743 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:31:29 +0000 (0:00:00.772)       0:28:53.515 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:31:30 +0000 (0:00:00.821)       0:28:54.337 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:31:31 +0000 (0:00:00.844)       0:28:55.181 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:31:32 +0000 (0:00:00.826)       0:28:56.008 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:31:33 +0000 (0:00:00.795)       0:28:56.804 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:31:34 +0000 (0:00:01.387)       0:28:58.192 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:31:36 +0000 (0:00:01.467)       0:28:59.659 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:31:37 +0000 (0:00:01.071)       0:29:00.731 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:31:37 +0000 (0:00:00.773)       0:29:01.504 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:31:38 +0000 (0:00:00.894)       0:29:02.399 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:31:40 +0000 (0:00:01.430)       0:29:03.829 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:31:41 +0000 (0:00:01.332)       0:29:05.162 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nThursday 26 March 2026  05:31:42 +0000 (0:00:01.152)       0:29:06.314 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nThursday 26 March 2026  05:31:44 +0000 (0:00:01.495)       0:29:07.809 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nThursday 26 March 2026  05:31:45 +0000 (0:00:01.158)       0:29:08.968 ******** \n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nThursday 26 March 2026  05:31:53 +0000 (0:00:07.848)       0:29:16.817 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:31:54 +0000 (0:00:01.260)       0:29:18.078 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:31:57 +0000 (0:00:03.250)       0:29:21.328 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nThursday 26 March 2026  05:31:59 +0000 (0:00:02.080)       0:29:23.409 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nThursday 26 March 2026  05:32:01 +0000 (0:00:01.496)       0:29:24.905 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:32:02 +0000 (0:00:00.755)       0:29:25.660 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:32:02 +0000 (0:00:00.758)       0:29:26.418 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:32:03 +0000 (0:00:01.114)       0:29:27.533 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nThursday 26 March 2026  05:32:04 +0000 (0:00:01.122)       0:29:28.655 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:32:06 +0000 (0:00:01.116)       0:29:29.772 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-1\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:32:07 +0000 (0:00:01.096)       0:29:30.868 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nThursday 26 March 2026  05:32:09 +0000 (0:00:02.028)       0:29:32.897 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nThursday 26 March 2026  05:32:11 +0000 (0:00:01.977)       0:29:34.874 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nThursday 26 March 2026  05:32:13 +0000 (0:00:02.354)       0:29:37.229 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:32:16 +0000 (0:00:03.301)       0:29:40.531 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Upgrade ceph mgr nodes] **************************************************\n\nTASK [Stop ceph mgr] ***********************************************************\nThursday 26 March 2026  05:32:17 +0000 (0:00:01.014)       0:29:41.546 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Mask ceph mgr systemd unit] **********************************************\nThursday 26 March 2026  05:32:20 +0000 (0:00:02.570)       0:29:44.116 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:32:22 +0000 (0:00:02.265)       0:29:46.381 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:32:23 +0000 (0:00:01.153)       0:29:47.534 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:32:25 +0000 (0:00:01.508)       0:29:49.042 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:32:26 +0000 (0:00:01.102)       0:29:50.145 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:32:27 +0000 (0:00:01.484)       0:29:51.630 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:32:29 +0000 (0:00:01.167)       0:29:52.798 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:32:30 +0000 (0:00:01.131)       0:29:53.930 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:32:31 +0000 (0:00:01.195)       0:29:55.125 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:32:32 +0000 (0:00:01.131)       0:29:56.256 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:32:33 +0000 (0:00:01.164)       0:29:57.421 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:32:35 +0000 (0:00:01.704)       0:29:59.126 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:32:36 +0000 (0:00:01.260)       0:30:00.386 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:32:39 +0000 (0:00:02.884)       0:30:03.271 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:32:41 +0000 (0:00:01.399)       0:30:04.671 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:32:43 +0000 (0:00:01.993)       0:30:06.665 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:32:44 +0000 (0:00:01.221)       0:30:07.887 ******** \n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:32:37.251347', 'end': '2026-03-26 05:32:37.306087', 'delta': '0:00:00.054740', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:32:37.856920', 'end': '2026-03-26 05:32:37.897986', 'delta': '0:00:00.041066', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:32:38.441829', 'end': '2026-03-26 05:32:38.485013', 'delta': '0:00:00.043184', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:32:45 +0000 (0:00:01.227)       0:30:09.114 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:32:46 +0000 (0:00:01.275)       0:30:10.389 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:32:48 +0000 (0:00:01.670)       0:30:12.060 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:32:49 +0000 (0:00:01.143)       0:30:13.204 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:32:51 +0000 (0:00:01.949)       0:30:15.154 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:32:52 +0000 (0:00:01.128)       0:30:16.282 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:32:53 +0000 (0:00:01.152)       0:30:17.435 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:32:55 +0000 (0:00:01.237)       0:30:18.673 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:32:56 +0000 (0:00:01.163)       0:30:19.836 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:32:57 +0000 (0:00:01.154)       0:30:20.991 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:32:58 +0000 (0:00:01.156)       0:30:22.148 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:32:59 +0000 (0:00:01.121)       0:30:23.270 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:33:00 +0000 (0:00:01.146)       0:30:24.416 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:33:01 +0000 (0:00:01.159)       0:30:25.576 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:33:03 +0000 (0:00:01.175)       0:30:26.752 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:33:04 +0000 (0:00:01.292)       0:30:28.045 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:33:05 +0000 (0:00:01.278)       0:30:29.323 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:33:07 +0000 (0:00:01.481)       0:30:30.804 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:33:08 +0000 (0:00:01.158)       0:30:31.962 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:33:09 +0000 (0:00:01.491)       0:30:33.454 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:33:10 +0000 (0:00:01.157)       0:30:34.612 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:33:12 +0000 (0:00:01.334)       0:30:35.947 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:33:13 +0000 (0:00:01.185)       0:30:37.132 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:33:15 +0000 (0:00:01.723)       0:30:38.856 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:33:16 +0000 (0:00:01.165)       0:30:40.022 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:33:17 +0000 (0:00:01.127)       0:30:41.150 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:33:19 +0000 (0:00:02.171)       0:30:43.322 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:33:21 +0000 (0:00:02.279)       0:30:45.601 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:33:23 +0000 (0:00:01.268)       0:30:46.870 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:33:24 +0000 (0:00:01.121)       0:30:47.991 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:33:25 +0000 (0:00:01.538)       0:30:49.530 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:33:27 +0000 (0:00:01.136)       0:30:50.667 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:33:28 +0000 (0:00:01.098)       0:30:51.765 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:33:29 +0000 (0:00:01.166)       0:30:52.932 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:33:30 +0000 (0:00:01.577)       0:30:54.509 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:33:32 +0000 (0:00:01.202)       0:30:55.712 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:33:33 +0000 (0:00:01.162)       0:30:56.875 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:33:34 +0000 (0:00:01.605)       0:30:58.480 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:33:36 +0000 (0:00:01.614)       0:31:00.095 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:33:37 +0000 (0:00:00.980)       0:31:01.076 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:33:38 +0000 (0:00:00.858)       0:31:01.934 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:33:39 +0000 (0:00:00.773)       0:31:02.707 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:33:39 +0000 (0:00:00.810)       0:31:03.518 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:33:40 +0000 (0:00:00.804)       0:31:04.322 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:33:41 +0000 (0:00:00.860)       0:31:05.183 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:33:42 +0000 (0:00:00.803)       0:31:05.986 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:33:43 +0000 (0:00:00.824)       0:31:06.811 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:33:44 +0000 (0:00:00.851)       0:31:07.662 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:33:44 +0000 (0:00:00.886)       0:31:08.549 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:33:45 +0000 (0:00:00.794)       0:31:09.343 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:33:46 +0000 (0:00:00.869)       0:31:10.213 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:33:47 +0000 (0:00:01.025)       0:31:11.239 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:33:48 +0000 (0:00:00.798)       0:31:12.038 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:33:49 +0000 (0:00:00.777)       0:31:12.816 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:33:49 +0000 (0:00:00.792)       0:31:13.608 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:33:50 +0000 (0:00:00.835)       0:31:14.443 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:33:51 +0000 (0:00:00.774)       0:31:15.218 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:33:52 +0000 (0:00:00.815)       0:31:16.034 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:33:53 +0000 (0:00:00.779)       0:31:16.813 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:33:53 +0000 (0:00:00.810)       0:31:17.623 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:33:54 +0000 (0:00:00.791)       0:31:18.415 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:33:56 +0000 (0:00:01.710)       0:31:20.125 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:33:58 +0000 (0:00:02.181)       0:31:22.307 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:34:00 +0000 (0:00:01.450)       0:31:23.758 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:34:01 +0000 (0:00:01.168)       0:31:24.926 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:34:02 +0000 (0:00:01.150)       0:31:26.077 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:34:04 +0000 (0:00:01.875)       0:31:27.952 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:34:05 +0000 (0:00:01.473)       0:31:29.426 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:34:06 +0000 (0:00:01.203)       0:31:30.629 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:34:07 +0000 (0:00:00.788)       0:31:31.417 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:34:08 +0000 (0:00:00.772)       0:31:32.190 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:34:09 +0000 (0:00:01.122)       0:31:33.313 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:34:11 +0000 (0:00:01.927)       0:31:35.241 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:34:12 +0000 (0:00:01.240)       0:31:36.481 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:34:14 +0000 (0:00:01.219)       0:31:37.700 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:34:15 +0000 (0:00:01.326)       0:31:39.027 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:34:16 +0000 (0:00:01.216)       0:31:40.244 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:34:17 +0000 (0:00:01.305)       0:31:41.550 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:34:18 +0000 (0:00:00.847)       0:31:42.398 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:34:21 +0000 (0:00:02.272)       0:31:44.670 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:34:21 +0000 (0:00:00.828)       0:31:45.498 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:34:23 +0000 (0:00:01.177)       0:31:46.676 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:34:24 +0000 (0:00:01.196)       0:31:47.873 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:34:25 +0000 (0:00:01.171)       0:31:49.044 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:34:26 +0000 (0:00:01.132)       0:31:50.177 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:34:27 +0000 (0:00:01.159)       0:31:51.336 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:34:28 +0000 (0:00:01.211)       0:31:52.548 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:34:30 +0000 (0:00:01.324)       0:31:53.873 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:34:31 +0000 (0:00:01.190)       0:31:55.063 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:34:32 +0000 (0:00:01.148)       0:31:56.212 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:34:33 +0000 (0:00:00.819)       0:31:57.032 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:34:34 +0000 (0:00:01.127)       0:31:58.159 ******** \n\u001b[0;32mok: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:34:40 +0000 (0:00:06.437)       0:32:04.596 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:34:41 +0000 (0:00:00.780)       0:32:05.377 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:34:42 +0000 (0:00:00.839)       0:32:06.217 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:34:43 +0000 (0:00:00.808)       0:32:07.025 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:34:44 +0000 (0:00:00.793)       0:32:07.819 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:34:44 +0000 (0:00:00.793)       0:32:08.613 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:34:45 +0000 (0:00:00.823)       0:32:09.436 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:34:46 +0000 (0:00:00.753)       0:32:10.190 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:34:47 +0000 (0:00:00.769)       0:32:10.959 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:34:48 +0000 (0:00:00.786)       0:32:11.745 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:34:48 +0000 (0:00:00.787)       0:32:12.533 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:34:49 +0000 (0:00:00.806)       0:32:13.340 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:34:50 +0000 (0:00:00.770)       0:32:14.110 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:34:51 +0000 (0:00:00.902)       0:32:15.013 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:34:52 +0000 (0:00:00.785)       0:32:15.798 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:34:53 +0000 (0:00:00.930)       0:32:16.729 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:34:53 +0000 (0:00:00.765)       0:32:17.495 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:34:54 +0000 (0:00:00.787)       0:32:18.282 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:34:55 +0000 (0:00:00.814)       0:32:19.097 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:34:56 +0000 (0:00:00.816)       0:32:19.914 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:34:57 +0000 (0:00:00.802)       0:32:20.716 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:34:57 +0000 (0:00:00.778)       0:32:21.495 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:34:58 +0000 (0:00:01.097)       0:32:22.592 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:34:59 +0000 (0:00:01.047)       0:32:23.639 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:35:01 +0000 (0:00:01.127)       0:32:24.767 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:35:01 +0000 (0:00:00.815)       0:32:25.582 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:35:02 +0000 (0:00:00.910)       0:32:26.493 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:35:04 +0000 (0:00:01.445)       0:32:27.938 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nThursday 26 March 2026  05:35:05 +0000 (0:00:01.653)       0:32:29.592 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nThursday 26 March 2026  05:35:07 +0000 (0:00:01.119)       0:32:30.712 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nThursday 26 March 2026  05:35:08 +0000 (0:00:01.472)       0:32:32.185 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nThursday 26 March 2026  05:35:09 +0000 (0:00:01.102)       0:32:33.287 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nThursday 26 March 2026  05:35:16 +0000 (0:00:07.050)       0:32:40.338 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:35:17 +0000 (0:00:01.191)       0:32:41.530 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:35:21 +0000 (0:00:03.340)       0:32:44.870 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nThursday 26 March 2026  05:35:23 +0000 (0:00:02.059)       0:32:46.929 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nThursday 26 March 2026  05:35:24 +0000 (0:00:01.520)       0:32:48.450 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nThursday 26 March 2026  05:35:25 +0000 (0:00:00.803)       0:32:49.253 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nThursday 26 March 2026  05:35:26 +0000 (0:00:00.794)       0:32:50.048 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:35:27 +0000 (0:00:01.099)       0:32:51.148 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nThursday 26 March 2026  05:35:28 +0000 (0:00:01.222)       0:32:52.371 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:35:29 +0000 (0:00:01.136)       0:32:53.507 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:35:31 +0000 (0:00:01.243)       0:32:54.751 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nThursday 26 March 2026  05:35:33 +0000 (0:00:02.052)       0:32:56.803 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nThursday 26 March 2026  05:35:35 +0000 (0:00:01.978)       0:32:58.782 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nThursday 26 March 2026  05:35:37 +0000 (0:00:02.506)       0:33:01.289 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nThursday 26 March 2026  05:35:41 +0000 (0:00:03.625)       0:33:04.914 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nThursday 26 March 2026  05:35:42 +0000 (0:00:01.571)       0:33:06.486 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nThursday 26 March 2026  05:35:45 +0000 (0:00:02.483)       0:33:08.970 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nThursday 26 March 2026  05:35:47 +0000 (0:00:02.344)       0:33:11.314 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nThursday 26 March 2026  05:35:48 +0000 (0:00:01.318)       0:33:12.632 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nThursday 26 March 2026  05:35:50 +0000 (0:00:01.156)       0:33:13.789 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=dashboard) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=prometheus) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nThursday 26 March 2026  05:35:51 +0000 (0:00:01.713)       0:33:15.502 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=dashboard) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=prometheus) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Set osd flags] ***********************************************************\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:35:53 +0000 (0:00:01.983)       0:33:17.486 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:35:55 +0000 (0:00:01.787)       0:33:19.274 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get pool list] ***********************************************************\nThursday 26 March 2026  05:35:57 +0000 (0:00:01.730)       0:33:21.004 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Get balancer module status] **********************************************\nThursday 26 March 2026  05:36:00 +0000 (0:00:02.915)       0:33:23.920 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Set_fact pools_pgautoscaler_mode] ****************************************\nThursday 26 March 2026  05:36:03 +0000 (0:00:02.829)       0:33:26.749 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 1, 'pool_name': '.mgr', 'create_time': '2026-03-26T02:57:09.996917+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 2, 'min_size': 1, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_acting': 6.059999942779541, 'score_stable': 6.059999942779541, 'optimal_score': 0.33000001311302185, 'raw_score_acting': 2, 'raw_score_stable': 2, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 2, 'pool_name': 'cephfs_data', 'create_time': '2026-03-26T02:58:21.868811+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '32', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '30', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'cephfs': {'data': 'cephfs'}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 3, 'pool_name': 'cephfs_metadata', 'create_time': '2026-03-26T02:58:25.656609+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 16, 'pg_placement_num': 16, 'pg_placement_num_target': 16, 'pg_num_target': 16, 'pg_num_pending': 16, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '64', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '30', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_autoscale_bias': 4, 'pg_num_min': 16, 'recovery_priority': 5}, 'application_metadata': {'cephfs': {'metadata': 'cephfs'}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 4, 'pool_name': 'default.rgw.buckets.data', 'create_time': '2026-03-26T02:59:25.200032+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '69', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 5, 'pool_name': 'default.rgw.buckets.index', 'create_time': '2026-03-26T02:59:31.330922+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '71', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 6, 'pool_name': 'default.rgw.control', 'create_time': '2026-03-26T02:59:37.488224+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '71', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.8799999952316284, 'score_stable': 1.8799999952316284, 'optimal_score': 1, 'raw_score_acting': 1.8799999952316284, 'raw_score_stable': 1.8799999952316284, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 7, 'pool_name': 'default.rgw.log', 'create_time': '2026-03-26T02:59:43.721154+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '194', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '73', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 8, 'pool_name': 'default.rgw.meta', 'create_time': '2026-03-26T02:59:50.241088+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '73', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 9, 'pool_name': '.rgw.root', 'create_time': '2026-03-26T03:00:02.559162+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 2, 'min_size': 1, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '75', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 10, 'pool_name': 'backups', 'create_time': '2026-03-26T03:00:50.660723+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '102', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 102, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 2.059999942779541, 'score_stable': 2.059999942779541, 'optimal_score': 1, 'raw_score_acting': 2.059999942779541, 'raw_score_stable': 2.059999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 11, 'pool_name': 'volumes', 'create_time': '2026-03-26T03:01:00.012041+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '109', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 109, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 12, 'pool_name': 'images', 'create_time': '2026-03-26T03:01:10.076355+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '204', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 6, 'snap_epoch': 204, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 13, 'pool_name': 'metrics', 'create_time': '2026-03-26T03:01:18.926005+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '125', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 125, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 14, 'pool_name': 'vms', 'create_time': '2026-03-26T03:01:27.854751+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '132', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 132, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m\n\nTASK [Disable balancer] ********************************************************\nThursday 26 March 2026  05:36:05 +0000 (0:00:02.826)       0:33:29.575 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Disable pg autoscale on pools] *******************************************\nThursday 26 March 2026  05:36:08 +0000 (0:00:03.027)       0:33:32.603 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.mgr', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_data', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_metadata', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.data', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.index', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.control', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.log', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.meta', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.rgw.root', 'mode': 'on'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'backups', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'volumes', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'images', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'metrics', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vms', 'mode': 'off'}) \u001b[0m\n\nTASK [Set osd flags] ***********************************************************\nThursday 26 March 2026  05:37:23 +0000 (0:01:14.542)       0:34:47.145 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=noout)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=nodeep-scrub)\u001b[0m\n\nPLAY [Upgrade ceph osds cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:37:29 +0000 (0:00:06.111)       0:34:53.257 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:37:30 +0000 (0:00:01.142)       0:34:54.399 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:37:32 +0000 (0:00:01.510)       0:34:55.910 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:37:33 +0000 (0:00:01.132)       0:34:57.043 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:37:34 +0000 (0:00:01.490)       0:34:58.534 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:37:36 +0000 (0:00:01.169)       0:34:59.703 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:37:37 +0000 (0:00:01.138)       0:35:00.842 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:37:38 +0000 (0:00:01.145)       0:35:01.987 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:37:39 +0000 (0:00:01.166)       0:35:03.153 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:37:40 +0000 (0:00:01.136)       0:35:04.290 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:37:42 +0000 (0:00:01.695)       0:35:05.986 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:37:43 +0000 (0:00:01.308)       0:35:07.294 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:37:46 +0000 (0:00:03.319)       0:35:10.613 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:37:48 +0000 (0:00:01.439)       0:35:12.053 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:37:50 +0000 (0:00:01.989)       0:35:14.043 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:37:51 +0000 (0:00:01.182)       0:35:15.226 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:37:44.202827', 'end': '2026-03-26 05:37:44.263426', 'delta': '0:00:00.060599', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:37:44.808591', 'end': '2026-03-26 05:37:44.855653', 'delta': '0:00:00.047062', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:37:45.729323', 'end': '2026-03-26 05:37:45.782176', 'delta': '0:00:00.052853', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:37:52 +0000 (0:00:01.196)       0:35:16.423 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:37:54 +0000 (0:00:01.672)       0:35:18.095 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:37:55 +0000 (0:00:01.266)       0:35:19.362 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:37:56 +0000 (0:00:01.164)       0:35:20.527 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:37:58 +0000 (0:00:01.979)       0:35:22.507 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:38:00 +0000 (0:00:01.155)       0:35:23.663 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:38:01 +0000 (0:00:01.146)       0:35:24.809 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:38:02 +0000 (0:00:01.225)       0:35:26.034 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:38:03 +0000 (0:00:01.121)       0:35:27.155 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:38:04 +0000 (0:00:01.114)       0:35:28.270 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:38:05 +0000 (0:00:01.199)       0:35:29.469 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:38:06 +0000 (0:00:01.152)       0:35:30.622 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:38:08 +0000 (0:00:01.200)       0:35:31.823 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:38:09 +0000 (0:00:01.118)       0:35:32.941 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:38:10 +0000 (0:00:01.135)       0:35:34.076 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:38:12 +0000 (0:00:01.740)       0:35:35.817 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:38:13 +0000 (0:00:01.360)       0:35:37.178 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:38:15 +0000 (0:00:01.523)       0:35:38.701 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:38:16 +0000 (0:00:01.114)       0:35:39.816 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:38:17 +0000 (0:00:01.470)       0:35:41.286 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:38:18 +0000 (0:00:01.100)       0:35:42.386 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:38:19 +0000 (0:00:01.257)       0:35:43.644 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:38:21 +0000 (0:00:01.126)       0:35:44.770 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:38:23 +0000 (0:00:02.083)       0:35:46.854 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:38:24 +0000 (0:00:01.240)       0:35:48.095 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:38:25 +0000 (0:00:01.138)       0:35:49.233 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:38:26 +0000 (0:00:01.171)       0:35:50.404 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:38:27 +0000 (0:00:01.135)       0:35:51.539 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:38:29 +0000 (0:00:01.169)       0:35:52.709 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:38:30 +0000 (0:00:01.234)       0:35:53.944 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:38:31 +0000 (0:00:01.411)       0:35:55.355 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:38:33 +0000 (0:00:01.424)       0:35:56.779 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:38:34 +0000 (0:00:01.373)       0:35:58.153 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:38:35 +0000 (0:00:01.143)       0:35:59.296 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:38:36 +0000 (0:00:01.352)       0:36:00.649 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:38:39 +0000 (0:00:02.178)       0:36:02.827 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get osd numbers - non container] *****************************************\nThursday 26 March 2026  05:38:41 +0000 (0:00:02.664)       0:36:05.492 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set num_osds] ************************************************************\nThursday 26 March 2026  05:38:43 +0000 (0:00:01.525)       0:36:07.018 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set_fact container_exec_cmd_osd] *****************************************\nThursday 26 March 2026  05:38:44 +0000 (0:00:01.180)       0:36:08.198 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Stop ceph osd] ***********************************************************\nThursday 26 March 2026  05:38:46 +0000 (0:00:01.655)       0:36:09.853 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:38:50 +0000 (0:00:04.297)       0:36:14.150 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:38:51 +0000 (0:00:01.179)       0:36:15.330 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:38:52 +0000 (0:00:01.143)       0:36:16.473 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:38:54 +0000 (0:00:01.211)       0:36:17.684 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:38:55 +0000 (0:00:01.518)       0:36:19.202 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:38:57 +0000 (0:00:01.483)       0:36:20.686 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:38:58 +0000 (0:00:01.503)       0:36:22.190 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:38:59 +0000 (0:00:01.219)       0:36:23.410 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:39:00 +0000 (0:00:01.172)       0:36:24.582 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:39:02 +0000 (0:00:01.125)       0:36:25.708 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:39:03 +0000 (0:00:01.514)       0:36:27.222 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:39:05 +0000 (0:00:01.505)       0:36:28.727 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:39:06 +0000 (0:00:01.160)       0:36:29.888 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:39:07 +0000 (0:00:01.140)       0:36:31.029 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:39:08 +0000 (0:00:01.157)       0:36:32.187 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:39:09 +0000 (0:00:01.181)       0:36:33.368 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:39:10 +0000 (0:00:01.148)       0:36:34.517 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:39:11 +0000 (0:00:01.102)       0:36:35.620 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:39:13 +0000 (0:00:01.140)       0:36:36.760 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:39:14 +0000 (0:00:01.137)       0:36:37.898 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:39:15 +0000 (0:00:01.200)       0:36:39.098 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:39:16 +0000 (0:00:01.166)       0:36:40.265 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:39:17 +0000 (0:00:01.212)       0:36:41.477 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:39:18 +0000 (0:00:01.155)       0:36:42.633 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:39:20 +0000 (0:00:01.115)       0:36:43.749 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:39:21 +0000 (0:00:01.125)       0:36:44.875 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:39:22 +0000 (0:00:01.151)       0:36:46.026 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:39:23 +0000 (0:00:01.174)       0:36:47.200 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:39:24 +0000 (0:00:01.102)       0:36:48.303 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:39:25 +0000 (0:00:01.124)       0:36:49.428 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:39:26 +0000 (0:00:01.159)       0:36:50.587 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:39:28 +0000 (0:00:01.134)       0:36:51.722 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:39:29 +0000 (0:00:01.177)       0:36:52.899 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:39:30 +0000 (0:00:01.104)       0:36:54.004 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:39:32 +0000 (0:00:01.971)       0:36:55.975 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:39:34 +0000 (0:00:02.286)       0:36:58.262 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:39:35 +0000 (0:00:01.153)       0:36:59.416 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:39:36 +0000 (0:00:01.125)       0:37:00.541 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:39:38 +0000 (0:00:01.155)       0:37:01.697 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:39:39 +0000 (0:00:01.805)       0:37:03.505 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:39:41 +0000 (0:00:01.514)       0:37:05.019 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:39:42 +0000 (0:00:01.228)       0:37:06.248 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:39:43 +0000 (0:00:01.235)       0:37:07.484 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:39:44 +0000 (0:00:01.158)       0:37:08.643 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:39:46 +0000 (0:00:01.111)       0:37:09.755 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:39:48 +0000 (0:00:02.889)       0:37:12.644 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:39:50 +0000 (0:00:01.176)       0:37:13.821 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:39:51 +0000 (0:00:01.179)       0:37:15.001 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:39:52 +0000 (0:00:01.146)       0:37:16.147 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:39:53 +0000 (0:00:01.167)       0:37:17.315 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:39:54 +0000 (0:00:01.178)       0:37:18.493 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:39:56 +0000 (0:00:01.183)       0:37:19.679 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:39:58 +0000 (0:00:02.581)       0:37:22.261 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:39:59 +0000 (0:00:01.108)       0:37:23.369 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:40:00 +0000 (0:00:01.098)       0:37:24.467 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:40:01 +0000 (0:00:01.142)       0:37:25.610 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:40:03 +0000 (0:00:01.203)       0:37:26.814 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:40:04 +0000 (0:00:01.138)       0:37:27.952 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:40:05 +0000 (0:00:01.184)       0:37:29.137 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:40:06 +0000 (0:00:01.131)       0:37:30.269 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:40:07 +0000 (0:00:01.130)       0:37:31.400 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:40:08 +0000 (0:00:01.231)       0:37:32.632 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:40:10 +0000 (0:00:01.102)       0:37:33.734 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:40:11 +0000 (0:00:01.175)       0:37:34.909 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:40:12 +0000 (0:00:01.146)       0:37:36.056 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:40:19 +0000 (0:00:06.674)       0:37:42.731 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  05:40:20 +0000 (0:00:01.506)       0:37:44.238 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  05:40:22 +0000 (0:00:01.533)       0:37:45.772 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:40:24 +0000 (0:00:02.002)       0:37:47.774 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:40:25 +0000 (0:00:01.145)       0:37:48.920 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:40:26 +0000 (0:00:01.112)       0:37:50.032 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:40:27 +0000 (0:00:01.206)       0:37:51.239 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:40:28 +0000 (0:00:01.179)       0:37:52.418 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:40:29 +0000 (0:00:01.159)       0:37:53.578 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:40:31 +0000 (0:00:01.197)       0:37:54.776 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:40:32 +0000 (0:00:01.126)       0:37:55.902 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:40:33 +0000 (0:00:01.176)       0:37:57.079 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:40:34 +0000 (0:00:01.154)       0:37:58.233 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:40:35 +0000 (0:00:01.136)       0:37:59.370 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:40:36 +0000 (0:00:01.263)       0:38:00.634 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:40:41 +0000 (0:00:04.522)       0:38:05.157 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:40:42 +0000 (0:00:01.245)       0:38:06.402 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:40:50 +0000 (0:00:08.009)       0:38:14.412 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:40:51 +0000 (0:00:01.139)       0:38:15.552 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:40:53 +0000 (0:00:01.147)       0:38:16.700 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:40:54 +0000 (0:00:01.178)       0:38:17.879 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:40:55 +0000 (0:00:01.214)       0:38:19.093 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:40:56 +0000 (0:00:01.137)       0:38:20.231 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:40:57 +0000 (0:00:01.307)       0:38:21.539 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:40:59 +0000 (0:00:01.448)       0:38:22.988 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:41:00 +0000 (0:00:01.419)       0:38:24.408 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:41:02 +0000 (0:00:01.516)       0:38:25.924 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:41:03 +0000 (0:00:01.181)       0:38:27.105 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:41:05 +0000 (0:00:01.916)       0:38:29.022 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nThursday 26 March 2026  05:41:07 +0000 (0:00:01.719)       0:38:30.741 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:41:08 +0000 (0:00:01.130)       0:38:31.871 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nThursday 26 March 2026  05:41:09 +0000 (0:00:01.668)       0:38:33.540 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nThursday 26 March 2026  05:41:11 +0000 (0:00:01.485)       0:38:35.026 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nThursday 26 March 2026  05:41:12 +0000 (0:00:01.114)       0:38:36.141 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nThursday 26 March 2026  05:41:13 +0000 (0:00:01.116)       0:38:37.258 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nThursday 26 March 2026  05:41:15 +0000 (0:00:01.434)       0:38:38.692 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nThursday 26 March 2026  05:41:16 +0000 (0:00:01.128)       0:38:39.820 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nThursday 26 March 2026  05:41:19 +0000 (0:00:02.994)       0:38:42.815 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nThursday 26 March 2026  05:41:20 +0000 (0:00:01.106)       0:38:43.922 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nThursday 26 March 2026  05:41:21 +0000 (0:00:01.566)       0:38:45.489 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:41:23 +0000 (0:00:01.826)       0:38:47.315 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:41:26 +0000 (0:00:03.222)       0:38:50.538 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nThursday 26 March 2026  05:41:28 +0000 (0:00:01.983)       0:38:52.522 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nThursday 26 March 2026  05:41:30 +0000 (0:00:01.247)       0:38:53.769 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nThursday 26 March 2026  05:41:31 +0000 (0:00:01.140)       0:38:54.910 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nThursday 26 March 2026  05:41:32 +0000 (0:00:01.191)       0:38:56.101 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nThursday 26 March 2026  05:41:33 +0000 (0:00:01.504)       0:38:57.606 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nThursday 26 March 2026  05:41:35 +0000 (0:00:01.458)       0:38:59.065 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:41:38 +0000 (0:00:03.593)       0:39:02.659 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:41:40 +0000 (0:00:01.484)       0:39:04.143 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nThursday 26 March 2026  05:41:42 +0000 (0:00:01.987)       0:39:06.130 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nThursday 26 March 2026  05:41:44 +0000 (0:00:01.964)       0:39:08.095 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:41:46 +0000 (0:00:02.264)       0:39:10.359 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nThursday 26 March 2026  05:41:47 +0000 (0:00:01.213)       0:39:11.573 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nThursday 26 March 2026  05:41:49 +0000 (0:00:01.120)       0:39:12.694 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nThursday 26 March 2026  05:41:50 +0000 (0:00:01.870)       0:39:14.564 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nThursday 26 March 2026  05:41:53 +0000 (0:00:02.955)       0:39:17.520 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nThursday 26 March 2026  05:41:58 +0000 (0:00:04.710)       0:39:22.230 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nThursday 26 March 2026  05:41:59 +0000 (0:00:01.260)       0:39:23.491 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nThursday 26 March 2026  05:42:01 +0000 (0:00:01.234)       0:39:24.726 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************\nThursday 26 March 2026  05:42:02 +0000 (0:00:01.270)       0:39:25.996 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***\nThursday 26 March 2026  05:42:03 +0000 (0:00:01.154)       0:39:27.150 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Waiting for clean pgs...] ************************************************\nThursday 26 March 2026  05:42:04 +0000 (0:00:01.204)       0:39:28.355 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (600 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (599 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (598 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (597 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:42:18 +0000 (0:00:14.164)       0:39:42.520 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:42:19 +0000 (0:00:01.121)       0:39:43.641 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:42:21 +0000 (0:00:01.130)       0:39:44.772 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:42:22 +0000 (0:00:01.156)       0:39:45.928 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:42:23 +0000 (0:00:01.144)       0:39:47.073 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:42:24 +0000 (0:00:01.141)       0:39:48.215 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:42:25 +0000 (0:00:01.107)       0:39:49.322 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Upgrade ceph osds cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:42:26 +0000 (0:00:00.953)       0:39:50.276 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:42:27 +0000 (0:00:01.167)       0:39:51.444 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:42:29 +0000 (0:00:01.481)       0:39:52.925 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:42:30 +0000 (0:00:01.166)       0:39:54.092 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:42:31 +0000 (0:00:01.452)       0:39:55.545 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:42:33 +0000 (0:00:01.180)       0:39:56.726 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:42:34 +0000 (0:00:01.124)       0:39:57.850 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:42:35 +0000 (0:00:01.144)       0:39:58.995 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:42:36 +0000 (0:00:01.137)       0:40:00.132 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:42:37 +0000 (0:00:01.117)       0:40:01.250 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:42:39 +0000 (0:00:02.101)       0:40:03.352 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:42:40 +0000 (0:00:01.253)       0:40:04.605 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:42:44 +0000 (0:00:03.297)       0:40:07.903 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:42:46 +0000 (0:00:01.828)       0:40:09.732 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:42:48 +0000 (0:00:01.963)       0:40:11.695 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:42:49 +0000 (0:00:01.239)       0:40:12.935 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:42:41.492746', 'end': '2026-03-26 05:42:41.541679', 'delta': '0:00:00.048933', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:42:42.438597', 'end': '2026-03-26 05:42:42.480128', 'delta': '0:00:00.041531', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:42:43.005995', 'end': '2026-03-26 05:42:43.056235', 'delta': '0:00:00.050240', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:42:50 +0000 (0:00:01.178)       0:40:14.113 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:42:51 +0000 (0:00:01.281)       0:40:15.394 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:42:53 +0000 (0:00:01.272)       0:40:16.667 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:42:54 +0000 (0:00:01.164)       0:40:17.831 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:42:56 +0000 (0:00:01.961)       0:40:19.793 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:42:57 +0000 (0:00:01.116)       0:40:20.910 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:42:58 +0000 (0:00:01.143)       0:40:22.053 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:42:59 +0000 (0:00:01.241)       0:40:23.294 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:43:00 +0000 (0:00:01.131)       0:40:24.426 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:43:01 +0000 (0:00:01.156)       0:40:25.583 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:43:03 +0000 (0:00:01.167)       0:40:26.751 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:43:04 +0000 (0:00:01.113)       0:40:27.865 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:43:05 +0000 (0:00:01.324)       0:40:29.190 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:43:06 +0000 (0:00:01.118)       0:40:30.308 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:43:07 +0000 (0:00:01.123)       0:40:31.431 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:43:09 +0000 (0:00:01.435)       0:40:32.866 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:43:10 +0000 (0:00:01.431)       0:40:34.298 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:43:12 +0000 (0:00:01.510)       0:40:35.808 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:43:13 +0000 (0:00:01.121)       0:40:36.930 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:43:14 +0000 (0:00:01.521)       0:40:38.451 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:43:15 +0000 (0:00:01.121)       0:40:39.573 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:43:17 +0000 (0:00:01.231)       0:40:40.805 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:43:18 +0000 (0:00:01.132)       0:40:41.938 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:43:20 +0000 (0:00:02.059)       0:40:43.997 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:43:21 +0000 (0:00:01.160)       0:40:45.158 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:43:22 +0000 (0:00:01.262)       0:40:46.421 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:43:23 +0000 (0:00:01.159)       0:40:47.581 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:43:25 +0000 (0:00:01.127)       0:40:48.708 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:43:26 +0000 (0:00:01.137)       0:40:49.845 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:43:27 +0000 (0:00:01.247)       0:40:51.093 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:43:28 +0000 (0:00:01.412)       0:40:52.506 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:43:30 +0000 (0:00:01.419)       0:40:53.925 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:43:31 +0000 (0:00:01.426)       0:40:55.352 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:43:32 +0000 (0:00:01.248)       0:40:56.600 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:43:34 +0000 (0:00:01.457)       0:40:58.057 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:43:36 +0000 (0:00:02.195)       0:41:00.253 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get osd numbers - non container] *****************************************\nThursday 26 March 2026  05:43:39 +0000 (0:00:02.419)       0:41:02.672 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set num_osds] ************************************************************\nThursday 26 March 2026  05:43:40 +0000 (0:00:01.127)       0:41:03.800 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set_fact container_exec_cmd_osd] *****************************************\nThursday 26 March 2026  05:43:40 +0000 (0:00:00.790)       0:41:04.591 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Stop ceph osd] ***********************************************************\nThursday 26 March 2026  05:43:41 +0000 (0:00:00.905)       0:41:05.496 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:43:45 +0000 (0:00:03.814)       0:41:09.311 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:43:46 +0000 (0:00:01.107)       0:41:10.419 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:43:47 +0000 (0:00:01.151)       0:41:11.570 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:43:49 +0000 (0:00:01.134)       0:41:12.704 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:43:50 +0000 (0:00:01.506)       0:41:14.211 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:43:52 +0000 (0:00:01.549)       0:41:15.761 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:43:53 +0000 (0:00:01.497)       0:41:17.258 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:43:54 +0000 (0:00:01.149)       0:41:18.408 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:43:55 +0000 (0:00:01.151)       0:41:19.560 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:43:57 +0000 (0:00:01.108)       0:41:20.668 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:43:58 +0000 (0:00:01.534)       0:41:22.202 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:44:00 +0000 (0:00:01.608)       0:41:23.811 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:44:00 +0000 (0:00:00.750)       0:41:24.562 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:44:01 +0000 (0:00:00.796)       0:41:25.359 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:44:02 +0000 (0:00:00.812)       0:41:26.171 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:44:03 +0000 (0:00:00.775)       0:41:26.947 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:44:04 +0000 (0:00:00.816)       0:41:27.763 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:44:04 +0000 (0:00:00.812)       0:41:28.576 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:44:05 +0000 (0:00:00.765)       0:41:29.341 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:44:06 +0000 (0:00:00.781)       0:41:30.123 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:44:07 +0000 (0:00:00.786)       0:41:30.910 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:44:08 +0000 (0:00:00.939)       0:41:31.849 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:44:09 +0000 (0:00:00.812)       0:41:32.662 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:44:09 +0000 (0:00:00.792)       0:41:33.454 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:44:10 +0000 (0:00:00.784)       0:41:34.239 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:44:11 +0000 (0:00:00.781)       0:41:35.020 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:44:12 +0000 (0:00:00.790)       0:41:35.810 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:44:12 +0000 (0:00:00.798)       0:41:36.609 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:44:13 +0000 (0:00:00.754)       0:41:37.363 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:44:14 +0000 (0:00:00.772)       0:41:38.136 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:44:15 +0000 (0:00:00.825)       0:41:38.962 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:44:16 +0000 (0:00:00.790)       0:41:39.752 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:44:16 +0000 (0:00:00.762)       0:41:40.515 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:44:17 +0000 (0:00:00.931)       0:41:41.447 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:44:19 +0000 (0:00:01.609)       0:41:43.057 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:44:21 +0000 (0:00:01.837)       0:41:44.894 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:44:22 +0000 (0:00:01.157)       0:41:46.052 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:44:23 +0000 (0:00:01.185)       0:41:47.237 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:44:24 +0000 (0:00:01.156)       0:41:48.394 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:44:26 +0000 (0:00:01.772)       0:41:50.167 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:44:27 +0000 (0:00:01.473)       0:41:51.640 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:44:29 +0000 (0:00:01.153)       0:41:52.793 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:44:29 +0000 (0:00:00.785)       0:41:53.580 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:44:30 +0000 (0:00:00.785)       0:41:54.365 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:44:32 +0000 (0:00:01.379)       0:41:55.744 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:44:33 +0000 (0:00:01.802)       0:41:57.547 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:44:35 +0000 (0:00:01.138)       0:41:58.685 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:44:36 +0000 (0:00:01.134)       0:41:59.820 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:44:37 +0000 (0:00:01.165)       0:42:00.985 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:44:38 +0000 (0:00:01.130)       0:42:02.116 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:44:39 +0000 (0:00:01.184)       0:42:03.301 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:44:40 +0000 (0:00:00.798)       0:42:04.099 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:44:42 +0000 (0:00:02.151)       0:42:06.251 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:44:43 +0000 (0:00:00.758)       0:42:07.009 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:44:44 +0000 (0:00:01.118)       0:42:08.128 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:44:45 +0000 (0:00:01.266)       0:42:09.394 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:44:46 +0000 (0:00:01.150)       0:42:10.545 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:44:48 +0000 (0:00:01.157)       0:42:11.702 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:44:49 +0000 (0:00:01.124)       0:42:12.827 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:44:50 +0000 (0:00:01.147)       0:42:13.974 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:44:51 +0000 (0:00:01.118)       0:42:15.093 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:44:52 +0000 (0:00:01.139)       0:42:16.233 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:44:53 +0000 (0:00:01.123)       0:42:17.356 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:44:54 +0000 (0:00:00.834)       0:42:18.191 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:44:55 +0000 (0:00:01.149)       0:42:19.340 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:45:01 +0000 (0:00:06.153)       0:42:25.494 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  05:45:03 +0000 (0:00:01.273)       0:42:26.768 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  05:45:04 +0000 (0:00:01.488)       0:42:28.257 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:45:06 +0000 (0:00:01.616)       0:42:29.874 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:45:06 +0000 (0:00:00.734)       0:42:30.608 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:45:07 +0000 (0:00:00.749)       0:42:31.357 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:45:08 +0000 (0:00:00.770)       0:42:32.128 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:45:09 +0000 (0:00:00.838)       0:42:32.967 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:45:10 +0000 (0:00:00.737)       0:42:33.704 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:45:10 +0000 (0:00:00.789)       0:42:34.493 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:45:11 +0000 (0:00:00.757)       0:42:35.250 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:45:12 +0000 (0:00:00.745)       0:42:35.996 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:45:13 +0000 (0:00:00.787)       0:42:36.784 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:45:13 +0000 (0:00:00.779)       0:42:37.563 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:45:14 +0000 (0:00:00.912)       0:42:38.476 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:45:18 +0000 (0:00:04.172)       0:42:42.648 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:45:19 +0000 (0:00:00.850)       0:42:43.498 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:45:27 +0000 (0:00:07.704)       0:42:51.203 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:45:28 +0000 (0:00:00.773)       0:42:51.976 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:45:29 +0000 (0:00:00.853)       0:42:52.829 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:45:29 +0000 (0:00:00.816)       0:42:53.646 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:45:30 +0000 (0:00:00.774)       0:42:54.421 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:45:31 +0000 (0:00:00.813)       0:42:55.235 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:45:32 +0000 (0:00:00.911)       0:42:56.147 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:45:33 +0000 (0:00:01.106)       0:42:57.253 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:45:35 +0000 (0:00:01.529)       0:42:58.782 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:45:36 +0000 (0:00:01.469)       0:43:00.252 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:45:37 +0000 (0:00:00.875)       0:43:01.128 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:45:38 +0000 (0:00:00.983)       0:43:02.111 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nThursday 26 March 2026  05:45:39 +0000 (0:00:01.388)       0:43:03.500 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:45:40 +0000 (0:00:00.778)       0:43:04.279 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nThursday 26 March 2026  05:45:41 +0000 (0:00:01.361)       0:43:05.640 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nThursday 26 March 2026  05:45:43 +0000 (0:00:01.123)       0:43:06.764 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nThursday 26 March 2026  05:45:44 +0000 (0:00:01.139)       0:43:07.904 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nThursday 26 March 2026  05:45:45 +0000 (0:00:01.177)       0:43:09.081 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nThursday 26 March 2026  05:45:46 +0000 (0:00:01.440)       0:43:10.522 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nThursday 26 March 2026  05:45:48 +0000 (0:00:01.161)       0:43:11.684 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nThursday 26 March 2026  05:45:50 +0000 (0:00:02.610)       0:43:14.295 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nThursday 26 March 2026  05:45:51 +0000 (0:00:00.776)       0:43:15.071 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nThursday 26 March 2026  05:45:52 +0000 (0:00:01.148)       0:43:16.220 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:45:54 +0000 (0:00:01.848)       0:43:18.068 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:45:57 +0000 (0:00:03.191)       0:43:21.260 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nThursday 26 March 2026  05:45:59 +0000 (0:00:01.582)       0:43:22.843 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nThursday 26 March 2026  05:46:00 +0000 (0:00:00.865)       0:43:23.708 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nThursday 26 March 2026  05:46:00 +0000 (0:00:00.802)       0:43:24.511 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nThursday 26 March 2026  05:46:01 +0000 (0:00:00.763)       0:43:25.275 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nThursday 26 March 2026  05:46:02 +0000 (0:00:01.133)       0:43:26.409 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nThursday 26 March 2026  05:46:04 +0000 (0:00:01.473)       0:43:27.882 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:46:07 +0000 (0:00:03.291)       0:43:31.174 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:46:08 +0000 (0:00:01.241)       0:43:32.416 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nThursday 26 March 2026  05:46:10 +0000 (0:00:01.995)       0:43:34.412 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nThursday 26 March 2026  05:46:12 +0000 (0:00:01.938)       0:43:36.350 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:46:14 +0000 (0:00:02.246)       0:43:38.597 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nThursday 26 March 2026  05:46:16 +0000 (0:00:01.201)       0:43:39.799 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nThursday 26 March 2026  05:46:17 +0000 (0:00:01.182)       0:43:40.981 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nThursday 26 March 2026  05:46:19 +0000 (0:00:01.812)       0:43:42.793 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nThursday 26 March 2026  05:46:22 +0000 (0:00:02.900)       0:43:45.694 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nThursday 26 March 2026  05:46:26 +0000 (0:00:04.206)       0:43:49.900 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nThursday 26 March 2026  05:46:27 +0000 (0:00:00.889)       0:43:50.790 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nThursday 26 March 2026  05:46:28 +0000 (0:00:00.885)       0:43:51.676 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************\nThursday 26 March 2026  05:46:29 +0000 (0:00:01.016)       0:43:52.693 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***\nThursday 26 March 2026  05:46:29 +0000 (0:00:00.750)       0:43:53.443 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Waiting for clean pgs...] ************************************************\nThursday 26 March 2026  05:46:30 +0000 (0:00:00.789)       0:43:54.232 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (600 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (599 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (598 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (597 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:46:44 +0000 (0:00:13.720)       0:44:07.953 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:46:45 +0000 (0:00:00.916)       0:44:08.870 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:46:46 +0000 (0:00:00.788)       0:44:09.659 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:46:46 +0000 (0:00:00.767)       0:44:10.426 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:46:47 +0000 (0:00:00.770)       0:44:11.197 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:46:48 +0000 (0:00:00.788)       0:44:11.985 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:46:49 +0000 (0:00:00.769)       0:44:12.755 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Upgrade ceph osds cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:46:50 +0000 (0:00:00.960)       0:44:13.715 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:46:51 +0000 (0:00:01.338)       0:44:15.053 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:46:52 +0000 (0:00:01.454)       0:44:16.508 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:46:53 +0000 (0:00:01.108)       0:44:17.617 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:46:55 +0000 (0:00:01.485)       0:44:19.102 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:46:56 +0000 (0:00:01.191)       0:44:20.294 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:46:57 +0000 (0:00:01.128)       0:44:21.423 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:46:58 +0000 (0:00:01.167)       0:44:22.590 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:47:00 +0000 (0:00:01.132)       0:44:23.723 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:47:01 +0000 (0:00:01.148)       0:44:24.871 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:47:03 +0000 (0:00:02.057)       0:44:26.929 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:47:04 +0000 (0:00:01.237)       0:44:28.166 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:47:07 +0000 (0:00:03.293)       0:44:31.460 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:47:09 +0000 (0:00:01.791)       0:44:33.252 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:47:11 +0000 (0:00:01.620)       0:44:34.872 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:47:12 +0000 (0:00:01.214)       0:44:36.087 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:47:05.394004', 'end': '2026-03-26 05:47:05.430712', 'delta': '0:00:00.036708', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:47:05.989006', 'end': '2026-03-26 05:47:06.037018', 'delta': '0:00:00.048012', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:47:06.596719', 'end': '2026-03-26 05:47:06.655213', 'delta': '0:00:00.058494', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:47:13 +0000 (0:00:01.178)       0:44:37.265 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:47:14 +0000 (0:00:01.258)       0:44:38.524 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:47:16 +0000 (0:00:01.293)       0:44:39.818 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:47:17 +0000 (0:00:01.139)       0:44:40.957 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:47:19 +0000 (0:00:02.028)       0:44:42.986 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:47:20 +0000 (0:00:01.319)       0:44:44.306 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:47:21 +0000 (0:00:01.155)       0:44:45.462 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:47:23 +0000 (0:00:01.251)       0:44:46.713 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:47:24 +0000 (0:00:01.141)       0:44:47.854 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:47:25 +0000 (0:00:01.154)       0:44:49.009 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:47:26 +0000 (0:00:01.234)       0:44:50.244 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:47:27 +0000 (0:00:01.124)       0:44:51.368 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:47:28 +0000 (0:00:01.176)       0:44:52.544 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:47:30 +0000 (0:00:01.146)       0:44:53.691 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:47:31 +0000 (0:00:01.186)       0:44:54.878 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:47:32 +0000 (0:00:01.422)       0:44:56.301 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:47:34 +0000 (0:00:01.415)       0:44:57.717 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:47:36 +0000 (0:00:02.516)       0:45:00.233 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:47:37 +0000 (0:00:01.138)       0:45:01.372 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:47:39 +0000 (0:00:01.502)       0:45:02.874 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:47:40 +0000 (0:00:01.151)       0:45:04.026 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:47:41 +0000 (0:00:01.240)       0:45:05.267 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:47:42 +0000 (0:00:01.173)       0:45:06.440 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:47:44 +0000 (0:00:02.128)       0:45:08.569 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:47:46 +0000 (0:00:01.204)       0:45:09.774 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:47:47 +0000 (0:00:01.160)       0:45:10.934 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:47:48 +0000 (0:00:01.141)       0:45:12.076 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:47:49 +0000 (0:00:01.115)       0:45:13.192 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:47:50 +0000 (0:00:01.168)       0:45:14.361 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:47:51 +0000 (0:00:01.257)       0:45:15.618 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:47:53 +0000 (0:00:01.387)       0:45:17.005 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:47:54 +0000 (0:00:01.362)       0:45:18.368 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:47:56 +0000 (0:00:01.417)       0:45:19.786 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:47:57 +0000 (0:00:01.184)       0:45:20.970 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:47:59 +0000 (0:00:01.756)       0:45:22.727 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:48:01 +0000 (0:00:02.170)       0:45:24.898 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Get osd numbers - non container] *****************************************\nThursday 26 March 2026  05:48:03 +0000 (0:00:02.264)       0:45:27.162 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set num_osds] ************************************************************\nThursday 26 March 2026  05:48:04 +0000 (0:00:01.150)       0:45:28.312 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set_fact container_exec_cmd_osd] *****************************************\nThursday 26 March 2026  05:48:05 +0000 (0:00:00.779)       0:45:29.092 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Stop ceph osd] ***********************************************************\nThursday 26 March 2026  05:48:06 +0000 (0:00:00.850)       0:45:29.943 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:48:10 +0000 (0:00:04.686)       0:45:34.630 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:48:12 +0000 (0:00:01.150)       0:45:35.780 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:48:13 +0000 (0:00:01.098)       0:45:36.879 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:48:14 +0000 (0:00:01.112)       0:45:37.991 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:48:15 +0000 (0:00:01.503)       0:45:39.495 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:48:17 +0000 (0:00:01.614)       0:45:41.110 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:48:18 +0000 (0:00:01.536)       0:45:42.646 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:48:20 +0000 (0:00:01.113)       0:45:43.760 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:48:21 +0000 (0:00:01.121)       0:45:44.882 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:48:22 +0000 (0:00:01.131)       0:45:46.014 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:48:23 +0000 (0:00:01.518)       0:45:47.533 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:48:25 +0000 (0:00:01.632)       0:45:49.165 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:48:26 +0000 (0:00:00.787)       0:45:49.952 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:48:27 +0000 (0:00:00.793)       0:45:50.746 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:48:27 +0000 (0:00:00.779)       0:45:51.525 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:48:28 +0000 (0:00:00.835)       0:45:52.361 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:48:29 +0000 (0:00:00.779)       0:45:53.140 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:48:30 +0000 (0:00:00.744)       0:45:53.885 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:48:30 +0000 (0:00:00.762)       0:45:54.647 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:48:31 +0000 (0:00:00.816)       0:45:55.463 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:48:32 +0000 (0:00:00.819)       0:45:56.282 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:48:33 +0000 (0:00:00.790)       0:45:57.073 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:48:34 +0000 (0:00:00.769)       0:45:57.842 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:48:34 +0000 (0:00:00.768)       0:45:58.611 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:48:35 +0000 (0:00:00.761)       0:45:59.373 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:48:36 +0000 (0:00:00.770)       0:46:00.144 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:48:37 +0000 (0:00:00.783)       0:46:00.928 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:48:38 +0000 (0:00:00.769)       0:46:01.698 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:48:38 +0000 (0:00:00.762)       0:46:02.461 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:48:39 +0000 (0:00:00.783)       0:46:03.244 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:48:40 +0000 (0:00:00.760)       0:46:04.005 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:48:41 +0000 (0:00:00.915)       0:46:04.920 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:48:42 +0000 (0:00:00.794)       0:46:05.715 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:48:42 +0000 (0:00:00.747)       0:46:06.463 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:48:44 +0000 (0:00:01.632)       0:46:08.095 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:48:46 +0000 (0:00:01.902)       0:46:09.998 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:48:47 +0000 (0:00:01.129)       0:46:11.127 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:48:48 +0000 (0:00:01.129)       0:46:12.256 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:48:49 +0000 (0:00:01.132)       0:46:13.389 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:48:51 +0000 (0:00:01.803)       0:46:15.193 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:48:53 +0000 (0:00:01.466)       0:46:16.659 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:48:54 +0000 (0:00:01.114)       0:46:17.774 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:48:55 +0000 (0:00:00.894)       0:46:18.668 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:48:55 +0000 (0:00:00.807)       0:46:19.476 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:48:56 +0000 (0:00:01.157)       0:46:20.633 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:48:58 +0000 (0:00:01.828)       0:46:22.463 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:48:59 +0000 (0:00:01.137)       0:46:23.601 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:49:01 +0000 (0:00:01.164)       0:46:24.765 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:49:02 +0000 (0:00:01.166)       0:46:25.932 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:49:03 +0000 (0:00:01.132)       0:46:27.064 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:49:04 +0000 (0:00:01.165)       0:46:28.230 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:49:05 +0000 (0:00:00.788)       0:46:29.018 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:49:07 +0000 (0:00:02.070)       0:46:31.088 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:49:08 +0000 (0:00:00.772)       0:46:31.861 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:49:09 +0000 (0:00:01.276)       0:46:33.137 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:49:10 +0000 (0:00:01.178)       0:46:34.315 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:49:11 +0000 (0:00:01.136)       0:46:35.452 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:49:12 +0000 (0:00:01.155)       0:46:36.608 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:49:14 +0000 (0:00:01.147)       0:46:37.756 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:49:15 +0000 (0:00:01.265)       0:46:39.021 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:49:16 +0000 (0:00:01.162)       0:46:40.184 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:49:17 +0000 (0:00:01.115)       0:46:41.300 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:49:18 +0000 (0:00:01.140)       0:46:42.440 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:49:19 +0000 (0:00:00.804)       0:46:43.244 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:49:20 +0000 (0:00:01.109)       0:46:44.354 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:49:27 +0000 (0:00:06.443)       0:46:50.798 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  05:49:28 +0000 (0:00:01.159)       0:46:51.957 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  05:49:29 +0000 (0:00:01.498)       0:46:53.456 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:49:31 +0000 (0:00:01.587)       0:46:55.044 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:49:32 +0000 (0:00:00.838)       0:46:55.883 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:49:33 +0000 (0:00:00.797)       0:46:56.681 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:49:33 +0000 (0:00:00.763)       0:46:57.444 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:49:34 +0000 (0:00:00.798)       0:46:58.242 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:49:35 +0000 (0:00:00.764)       0:46:59.007 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:49:36 +0000 (0:00:00.757)       0:46:59.765 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:49:36 +0000 (0:00:00.750)       0:47:00.515 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:49:37 +0000 (0:00:00.771)       0:47:01.287 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:49:38 +0000 (0:00:00.840)       0:47:02.128 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:49:39 +0000 (0:00:00.786)       0:47:02.915 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:49:40 +0000 (0:00:00.848)       0:47:03.763 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:49:44 +0000 (0:00:04.124)       0:47:07.888 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:49:45 +0000 (0:00:00.850)       0:47:08.738 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:49:52 +0000 (0:00:07.356)       0:47:16.095 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:49:53 +0000 (0:00:00.790)       0:47:16.885 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:49:54 +0000 (0:00:00.790)       0:47:17.676 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:49:54 +0000 (0:00:00.824)       0:47:18.501 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:49:55 +0000 (0:00:00.816)       0:47:19.318 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:49:56 +0000 (0:00:00.794)       0:47:20.113 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:49:57 +0000 (0:00:00.932)       0:47:21.045 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:49:58 +0000 (0:00:01.433)       0:47:22.478 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:50:00 +0000 (0:00:01.487)       0:47:23.966 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:50:01 +0000 (0:00:01.081)       0:47:25.048 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:50:02 +0000 (0:00:00.805)       0:47:25.854 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:50:03 +0000 (0:00:01.035)       0:47:26.889 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nThursday 26 March 2026  05:50:04 +0000 (0:00:01.481)       0:47:28.370 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nThursday 26 March 2026  05:50:05 +0000 (0:00:00.794)       0:47:29.165 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nThursday 26 March 2026  05:50:06 +0000 (0:00:01.470)       0:47:30.635 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nThursday 26 March 2026  05:50:08 +0000 (0:00:01.062)       0:47:31.698 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nThursday 26 March 2026  05:50:09 +0000 (0:00:01.124)       0:47:32.822 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nThursday 26 March 2026  05:50:10 +0000 (0:00:01.083)       0:47:33.905 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nThursday 26 March 2026  05:50:11 +0000 (0:00:01.468)       0:47:35.374 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nThursday 26 March 2026  05:50:12 +0000 (0:00:01.108)       0:47:36.482 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nThursday 26 March 2026  05:50:16 +0000 (0:00:03.633)       0:47:40.116 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nThursday 26 March 2026  05:50:17 +0000 (0:00:00.765)       0:47:40.882 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nThursday 26 March 2026  05:50:18 +0000 (0:00:01.113)       0:47:41.995 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:50:20 +0000 (0:00:01.798)       0:47:43.794 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:50:23 +0000 (0:00:03.176)       0:47:46.971 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nThursday 26 March 2026  05:50:24 +0000 (0:00:01.623)       0:47:48.594 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nThursday 26 March 2026  05:50:25 +0000 (0:00:00.906)       0:47:49.501 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nThursday 26 March 2026  05:50:26 +0000 (0:00:00.770)       0:47:50.271 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nThursday 26 March 2026  05:50:27 +0000 (0:00:00.748)       0:47:51.020 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nThursday 26 March 2026  05:50:28 +0000 (0:00:01.235)       0:47:52.255 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nThursday 26 March 2026  05:50:30 +0000 (0:00:01.465)       0:47:53.721 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:50:33 +0000 (0:00:03.351)       0:47:57.072 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:50:34 +0000 (0:00:01.106)       0:47:58.179 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nThursday 26 March 2026  05:50:36 +0000 (0:00:01.966)       0:48:00.145 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nThursday 26 March 2026  05:50:38 +0000 (0:00:01.927)       0:48:02.073 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nThursday 26 March 2026  05:50:40 +0000 (0:00:02.255)       0:48:04.329 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nThursday 26 March 2026  05:50:41 +0000 (0:00:01.120)       0:48:05.449 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nThursday 26 March 2026  05:50:42 +0000 (0:00:01.154)       0:48:06.604 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nThursday 26 March 2026  05:50:44 +0000 (0:00:01.848)       0:48:08.452 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nThursday 26 March 2026  05:50:47 +0000 (0:00:02.863)       0:48:11.315 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nThursday 26 March 2026  05:50:51 +0000 (0:00:04.175)       0:48:15.491 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nThursday 26 March 2026  05:50:52 +0000 (0:00:00.909)       0:48:16.401 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nThursday 26 March 2026  05:51:06 +0000 (0:00:13.348)       0:48:29.749 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************\nThursday 26 March 2026  05:51:06 +0000 (0:00:00.882)       0:48:30.632 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***\nThursday 26 March 2026  05:51:07 +0000 (0:00:00.749)       0:48:31.382 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Waiting for clean pgs...] ************************************************\nThursday 26 March 2026  05:51:08 +0000 (0:00:00.787)       0:48:32.169 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nThursday 26 March 2026  05:51:10 +0000 (0:00:01.931)       0:48:34.101 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nThursday 26 March 2026  05:51:11 +0000 (0:00:00.779)       0:48:34.880 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nThursday 26 March 2026  05:51:12 +0000 (0:00:00.814)       0:48:35.695 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nThursday 26 March 2026  05:51:12 +0000 (0:00:00.795)       0:48:36.491 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nThursday 26 March 2026  05:51:13 +0000 (0:00:00.776)       0:48:37.268 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nThursday 26 March 2026  05:51:14 +0000 (0:00:00.769)       0:48:38.037 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nThursday 26 March 2026  05:51:15 +0000 (0:00:00.764)       0:48:38.802 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Complete osd upgrade] ****************************************************\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:51:16 +0000 (0:00:01.800)       0:48:40.602 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:51:18 +0000 (0:00:01.868)       0:48:42.471 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Re-enable pg autoscale on pools] *****************************************\nThursday 26 March 2026  05:51:20 +0000 (0:00:01.358)       0:48:43.830 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.mgr', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_data', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_metadata', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.data', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.index', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.control', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.log', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.meta', 'mode': 'on'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.rgw.root', 'mode': 'on'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'backups', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'volumes', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'images', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'metrics', 'mode': 'off'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vms', 'mode': 'off'}) \u001b[0m\n\nTASK [Unset osd flags] *********************************************************\nThursday 26 March 2026  05:52:34 +0000 (0:01:14.512)       0:49:58.342 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=noout)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=nodeep-scrub)\u001b[0m\n\nTASK [Re-enable balancer] ******************************************************\nThursday 26 March 2026  05:52:41 +0000 (0:00:06.923)       0:50:05.265 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nPLAY [Upgrade ceph mdss cluster, deactivate all rank > 0] **********************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:52:44 +0000 (0:00:03.174)       0:50:08.439 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:52:45 +0000 (0:00:01.181)       0:50:09.621 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:52:47 +0000 (0:00:01.579)       0:50:11.201 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:52:48 +0000 (0:00:01.207)       0:50:12.409 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:52:50 +0000 (0:00:01.523)       0:50:13.932 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:52:51 +0000 (0:00:01.180)       0:50:15.113 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:52:52 +0000 (0:00:01.197)       0:50:16.310 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:52:53 +0000 (0:00:01.188)       0:50:17.499 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:52:55 +0000 (0:00:01.166)       0:50:18.665 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:52:56 +0000 (0:00:01.243)       0:50:19.909 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:52:57 +0000 (0:00:01.682)       0:50:21.592 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:52:59 +0000 (0:00:01.300)       0:50:22.893 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:53:02 +0000 (0:00:03.160)       0:50:26.053 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:53:03 +0000 (0:00:01.490)       0:50:27.543 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:53:05 +0000 (0:00:01.726)       0:50:29.270 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:53:06 +0000 (0:00:01.215)       0:50:30.486 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:52:59.842534', 'end': '2026-03-26 05:52:59.895695', 'delta': '0:00:00.053161', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:53:00.541458', 'end': '2026-03-26 05:53:00.583420', 'delta': '0:00:00.041962', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:53:01.105903', 'end': '2026-03-26 05:53:01.153302', 'delta': '0:00:00.047399', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:53:08 +0000 (0:00:01.263)       0:50:31.749 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:53:09 +0000 (0:00:01.304)       0:50:33.054 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:53:10 +0000 (0:00:01.316)       0:50:34.371 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:53:11 +0000 (0:00:01.155)       0:50:35.526 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:53:13 +0000 (0:00:02.035)       0:50:37.561 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:53:15 +0000 (0:00:01.161)       0:50:38.723 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:53:16 +0000 (0:00:01.250)       0:50:39.974 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:53:17 +0000 (0:00:01.273)       0:50:41.247 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:53:18 +0000 (0:00:01.155)       0:50:42.402 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:53:19 +0000 (0:00:01.147)       0:50:43.550 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:53:21 +0000 (0:00:01.180)       0:50:44.731 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:53:22 +0000 (0:00:01.190)       0:50:45.921 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:53:23 +0000 (0:00:01.132)       0:50:47.054 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:53:24 +0000 (0:00:01.180)       0:50:48.235 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:53:25 +0000 (0:00:01.166)       0:50:49.402 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:53:27 +0000 (0:00:01.310)       0:50:50.712 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:53:28 +0000 (0:00:01.342)       0:50:52.055 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:53:29 +0000 (0:00:01.543)       0:50:53.599 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:53:31 +0000 (0:00:01.182)       0:50:54.781 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:53:32 +0000 (0:00:01.520)       0:50:56.302 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:53:33 +0000 (0:00:01.181)       0:50:57.483 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:53:35 +0000 (0:00:01.248)       0:50:58.732 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:53:36 +0000 (0:00:01.141)       0:50:59.873 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:53:37 +0000 (0:00:01.714)       0:51:01.588 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:53:39 +0000 (0:00:01.198)       0:51:02.786 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:53:40 +0000 (0:00:01.157)       0:51:03.944 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:53:42 +0000 (0:00:02.213)       0:51:06.157 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Set max_mds 1 on ceph fs] ************************************************\nThursday 26 March 2026  05:53:45 +0000 (0:00:02.776)       0:51:08.933 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Wait until only rank 0 is up] ********************************************\nThursday 26 March 2026  05:53:48 +0000 (0:00:03.190)       0:51:12.124 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get name of remaining active mds] ****************************************\nThursday 26 March 2026  05:53:51 +0000 (0:00:02.985)       0:51:15.110 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set_fact mds_active_name] ************************************************\nThursday 26 March 2026  05:53:53 +0000 (0:00:02.115)       0:51:17.225 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'gid_4752', 'value': {'gid': 4752, 'name': 'testbed-node-3', 'rank': 0, 'incarnation': 4, 'state': 'up:active', 'state_seq': 2, 'addr': '192.168.16.13:6817/3459244977', 'addrs': {'addrvec': [{'type': 'v2', 'addr': '192.168.16.13:6816', 'nonce': 3459244977}, {'type': 'v1', 'addr': '192.168.16.13:6817', 'nonce': 3459244977}]}, 'join_fscid': -1, 'export_targets': [], 'features': 4540138322906710015, 'flags': 0, 'compat': {'compat': {}, 'ro_compat': {}, 'incompat': {'feature_1': 'base v0.20', 'feature_2': 'client writeable ranges', 'feature_3': 'default file layouts on dirs', 'feature_4': 'dir inode in separate object', 'feature_5': 'mds uses versioned encoding', 'feature_6': 'dirfrag is stored in omap', 'feature_7': 'mds uses inline data', 'feature_8': 'no anchor table', 'feature_9': 'file layout v2', 'feature_10': 'snaprealm v2'}}}})\u001b[0m\n\nTASK [Set_fact mds_active_host] ************************************************\nThursday 26 March 2026  05:53:54 +0000 (0:00:01.240)       0:51:18.466 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-3)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\nTASK [Create standby_mdss group] ***********************************************\nThursday 26 March 2026  05:53:56 +0000 (0:00:01.553)       0:51:20.020 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-5)\u001b[0m\n\nTASK [Stop standby ceph mds] ***************************************************\nThursday 26 March 2026  05:53:57 +0000 (0:00:01.559)       0:51:21.579 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\nTASK [Mask systemd units for standby ceph mds] *********************************\nThursday 26 March 2026  05:54:09 +0000 (0:00:11.167)       0:51:32.747 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\nTASK [Wait until all standbys mds are stopped] *********************************\nThursday 26 March 2026  05:54:13 +0000 (0:00:04.755)       0:51:37.502 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Create active_mdss group] ************************************************\nThursday 26 March 2026  05:54:16 +0000 (0:00:02.186)       0:51:39.688 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Upgrade active mds] ******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:54:17 +0000 (0:00:01.601)       0:51:41.290 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:54:18 +0000 (0:00:01.144)       0:51:42.434 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:54:20 +0000 (0:00:01.438)       0:51:43.873 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:54:21 +0000 (0:00:01.110)       0:51:44.983 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:54:22 +0000 (0:00:01.428)       0:51:46.412 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:54:23 +0000 (0:00:01.135)       0:51:47.548 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:54:25 +0000 (0:00:01.113)       0:51:48.661 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:54:26 +0000 (0:00:01.181)       0:51:49.842 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:54:27 +0000 (0:00:01.134)       0:51:50.977 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:54:28 +0000 (0:00:01.129)       0:51:52.107 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:54:30 +0000 (0:00:01.751)       0:51:53.858 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:54:31 +0000 (0:00:01.273)       0:51:55.131 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:54:34 +0000 (0:00:02.843)       0:51:57.975 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:54:35 +0000 (0:00:01.507)       0:51:59.482 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:54:37 +0000 (0:00:01.981)       0:52:01.464 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:54:38 +0000 (0:00:01.153)       0:52:02.618 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:54:32.014467', 'end': '2026-03-26 05:54:32.067837', 'delta': '0:00:00.053370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:54:32.555012', 'end': '2026-03-26 05:54:32.590382', 'delta': '0:00:00.035370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:54:33.112871', 'end': '2026-03-26 05:54:33.160580', 'delta': '0:00:00.047709', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:54:40 +0000 (0:00:01.208)       0:52:03.826 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:54:41 +0000 (0:00:01.304)       0:52:05.131 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:54:43 +0000 (0:00:01.662)       0:52:06.793 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:54:44 +0000 (0:00:01.149)       0:52:07.942 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:54:46 +0000 (0:00:02.011)       0:52:09.954 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:54:47 +0000 (0:00:01.248)       0:52:11.204 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:54:48 +0000 (0:00:01.161)       0:52:12.365 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:54:49 +0000 (0:00:01.263)       0:52:13.629 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:54:51 +0000 (0:00:01.166)       0:52:14.795 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:54:52 +0000 (0:00:01.125)       0:52:15.921 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:54:53 +0000 (0:00:01.166)       0:52:17.087 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:54:54 +0000 (0:00:01.114)       0:52:18.201 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:54:55 +0000 (0:00:01.226)       0:52:19.428 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:54:56 +0000 (0:00:01.129)       0:52:20.557 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:54:58 +0000 (0:00:01.171)       0:52:21.729 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:54:59 +0000 (0:00:01.516)       0:52:23.246 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:55:01 +0000 (0:00:01.431)       0:52:24.678 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:55:02 +0000 (0:00:01.579)       0:52:26.257 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:55:03 +0000 (0:00:01.132)       0:52:27.390 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:55:05 +0000 (0:00:01.478)       0:52:28.869 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:55:06 +0000 (0:00:01.211)       0:52:30.081 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:55:07 +0000 (0:00:01.192)       0:52:31.273 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:55:08 +0000 (0:00:01.107)       0:52:32.381 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:55:10 +0000 (0:00:01.840)       0:52:34.221 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:55:11 +0000 (0:00:01.134)       0:52:35.356 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:55:12 +0000 (0:00:01.111)       0:52:36.468 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:55:13 +0000 (0:00:01.174)       0:52:37.642 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:55:15 +0000 (0:00:01.347)       0:52:38.990 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:55:16 +0000 (0:00:01.161)       0:52:40.152 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:55:17 +0000 (0:00:01.327)       0:52:41.479 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:55:19 +0000 (0:00:01.450)       0:52:42.930 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:55:20 +0000 (0:00:01.404)       0:52:44.335 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:55:22 +0000 (0:00:01.468)       0:52:45.803 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:55:23 +0000 (0:00:01.194)       0:52:46.998 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:55:24 +0000 (0:00:01.338)       0:52:48.336 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:55:26 +0000 (0:00:02.312)       0:52:50.649 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Prevent restart from the packaging] **************************************\nThursday 26 March 2026  05:55:29 +0000 (0:00:02.663)       0:52:53.312 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:55:30 +0000 (0:00:01.132)       0:52:54.445 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:55:31 +0000 (0:00:01.144)       0:52:55.589 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:55:33 +0000 (0:00:01.308)       0:52:56.898 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:55:34 +0000 (0:00:01.127)       0:52:58.025 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:55:35 +0000 (0:00:01.504)       0:52:59.530 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:55:37 +0000 (0:00:01.562)       0:53:01.093 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:55:38 +0000 (0:00:01.552)       0:53:02.646 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:55:40 +0000 (0:00:01.131)       0:53:03.778 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:55:41 +0000 (0:00:01.194)       0:53:04.972 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:55:42 +0000 (0:00:01.121)       0:53:06.094 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:55:44 +0000 (0:00:01.588)       0:53:07.682 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:55:45 +0000 (0:00:01.590)       0:53:09.272 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:55:46 +0000 (0:00:01.158)       0:53:10.431 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:55:47 +0000 (0:00:01.226)       0:53:11.657 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:55:49 +0000 (0:00:01.186)       0:53:12.844 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:55:50 +0000 (0:00:01.126)       0:53:13.971 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:55:51 +0000 (0:00:01.182)       0:53:15.153 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:55:52 +0000 (0:00:01.164)       0:53:16.317 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  05:55:53 +0000 (0:00:01.201)       0:53:17.518 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  05:55:55 +0000 (0:00:01.159)       0:53:18.678 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  05:55:56 +0000 (0:00:01.157)       0:53:19.836 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  05:55:57 +0000 (0:00:01.175)       0:53:21.011 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  05:55:58 +0000 (0:00:01.145)       0:53:22.157 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  05:55:59 +0000 (0:00:01.151)       0:53:23.309 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  05:56:00 +0000 (0:00:01.158)       0:53:24.468 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  05:56:01 +0000 (0:00:01.132)       0:53:25.601 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  05:56:03 +0000 (0:00:01.237)       0:53:26.838 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  05:56:04 +0000 (0:00:01.145)       0:53:27.983 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  05:56:05 +0000 (0:00:01.191)       0:53:29.175 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  05:56:06 +0000 (0:00:01.141)       0:53:30.317 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  05:56:07 +0000 (0:00:01.202)       0:53:31.519 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  05:56:09 +0000 (0:00:01.180)       0:53:32.699 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  05:56:10 +0000 (0:00:01.135)       0:53:33.835 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  05:56:11 +0000 (0:00:01.128)       0:53:34.964 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  05:56:13 +0000 (0:00:01.889)       0:53:36.853 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  05:56:15 +0000 (0:00:02.250)       0:53:39.104 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  05:56:16 +0000 (0:00:01.120)       0:53:40.224 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  05:56:17 +0000 (0:00:01.154)       0:53:41.378 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  05:56:19 +0000 (0:00:01.304)       0:53:42.682 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  05:56:20 +0000 (0:00:01.869)       0:53:44.553 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  05:56:22 +0000 (0:00:01.497)       0:53:46.051 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  05:56:23 +0000 (0:00:01.144)       0:53:47.196 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  05:56:24 +0000 (0:00:01.155)       0:53:48.351 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  05:56:25 +0000 (0:00:01.180)       0:53:49.532 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  05:56:27 +0000 (0:00:01.190)       0:53:50.723 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  05:56:28 +0000 (0:00:01.815)       0:53:52.538 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  05:56:30 +0000 (0:00:01.166)       0:53:53.705 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  05:56:31 +0000 (0:00:01.166)       0:53:54.871 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  05:56:32 +0000 (0:00:01.187)       0:53:56.059 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  05:56:33 +0000 (0:00:01.199)       0:53:57.258 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  05:56:34 +0000 (0:00:01.391)       0:53:58.650 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  05:56:36 +0000 (0:00:01.196)       0:53:59.847 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  05:56:38 +0000 (0:00:02.550)       0:54:02.398 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  05:56:39 +0000 (0:00:01.135)       0:54:03.533 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  05:56:41 +0000 (0:00:01.155)       0:54:04.689 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  05:56:42 +0000 (0:00:01.175)       0:54:05.865 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  05:56:43 +0000 (0:00:01.150)       0:54:07.016 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  05:56:44 +0000 (0:00:01.146)       0:54:08.162 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  05:56:45 +0000 (0:00:01.101)       0:54:09.264 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  05:56:46 +0000 (0:00:01.177)       0:54:10.442 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  05:56:47 +0000 (0:00:01.186)       0:54:11.629 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  05:56:49 +0000 (0:00:01.124)       0:54:12.754 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  05:56:50 +0000 (0:00:01.228)       0:54:13.982 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  05:56:51 +0000 (0:00:01.151)       0:54:15.133 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  05:56:52 +0000 (0:00:01.127)       0:54:16.261 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  05:56:59 +0000 (0:00:06.553)       0:54:22.815 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  05:57:00 +0000 (0:00:01.161)       0:54:23.976 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  05:57:01 +0000 (0:00:01.498)       0:54:25.474 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  05:57:04 +0000 (0:00:03.000)       0:54:28.475 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  05:57:05 +0000 (0:00:01.122)       0:54:29.598 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  05:57:07 +0000 (0:00:01.118)       0:54:30.717 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  05:57:08 +0000 (0:00:01.153)       0:54:31.871 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  05:57:09 +0000 (0:00:01.214)       0:54:33.085 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  05:57:10 +0000 (0:00:01.139)       0:54:34.224 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  05:57:11 +0000 (0:00:01.126)       0:54:35.351 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  05:57:12 +0000 (0:00:01.113)       0:54:36.465 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  05:57:13 +0000 (0:00:01.106)       0:54:37.572 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  05:57:15 +0000 (0:00:01.128)       0:54:38.701 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  05:57:16 +0000 (0:00:01.172)       0:54:39.873 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  05:57:17 +0000 (0:00:01.211)       0:54:41.085 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  05:57:21 +0000 (0:00:04.492)       0:54:45.578 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  05:57:23 +0000 (0:00:01.201)       0:54:46.779 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  05:57:27 +0000 (0:00:04.801)       0:54:51.581 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  05:57:29 +0000 (0:00:01.133)       0:54:52.714 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:57:30 +0000 (0:00:01.110)       0:54:53.825 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:57:31 +0000 (0:00:01.222)       0:54:55.048 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:57:32 +0000 (0:00:01.180)       0:54:56.228 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:57:33 +0000 (0:00:01.184)       0:54:57.413 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:57:34 +0000 (0:00:01.233)       0:54:58.647 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:57:36 +0000 (0:00:01.520)       0:55:00.168 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:57:37 +0000 (0:00:01.450)       0:55:01.619 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:57:39 +0000 (0:00:01.470)       0:55:03.089 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:57:40 +0000 (0:00:01.175)       0:55:04.265 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  05:57:41 +0000 (0:00:01.384)       0:55:05.650 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nThursday 26 March 2026  05:57:43 +0000 (0:00:01.845)       0:55:07.495 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nThursday 26 March 2026  05:57:44 +0000 (0:00:01.154)       0:55:08.649 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nThursday 26 March 2026  05:57:46 +0000 (0:00:01.819)       0:55:10.469 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nThursday 26 March 2026  05:57:48 +0000 (0:00:01.987)       0:55:12.457 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  05:57:51 +0000 (0:00:03.181)       0:55:15.639 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nThursday 26 March 2026  05:57:53 +0000 (0:00:01.946)       0:55:17.585 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nThursday 26 March 2026  05:57:55 +0000 (0:00:01.579)       0:55:19.165 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nThursday 26 March 2026  05:57:56 +0000 (0:00:01.149)       0:55:20.314 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  05:57:58 +0000 (0:00:01.552)       0:55:21.867 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nThursday 26 March 2026  05:57:59 +0000 (0:00:01.576)       0:55:23.444 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nThursday 26 March 2026  05:58:01 +0000 (0:00:01.996)       0:55:25.440 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nThursday 26 March 2026  05:58:03 +0000 (0:00:01.894)       0:55:27.335 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nThursday 26 March 2026  05:58:05 +0000 (0:00:02.290)       0:55:29.625 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nThursday 26 March 2026  05:58:08 +0000 (0:00:02.324)       0:55:31.950 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Restart ceph mds] ********************************************************\nThursday 26 March 2026  05:58:09 +0000 (0:00:01.635)       0:55:33.586 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Restart active mds] ******************************************************\nThursday 26 March 2026  05:58:11 +0000 (0:00:01.165)       0:55:34.751 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Upgrade standbys ceph mdss cluster] **************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  05:58:15 +0000 (0:00:04.539)       0:55:39.291 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  05:58:17 +0000 (0:00:01.442)       0:55:40.734 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  05:58:18 +0000 (0:00:01.673)       0:55:42.407 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  05:58:20 +0000 (0:00:01.527)       0:55:43.936 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  05:58:21 +0000 (0:00:01.530)       0:55:45.467 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  05:58:23 +0000 (0:00:01.410)       0:55:46.878 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  05:58:24 +0000 (0:00:01.338)       0:55:48.217 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  05:58:25 +0000 (0:00:01.342)       0:55:49.559 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  05:58:27 +0000 (0:00:01.272)       0:55:50.832 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  05:58:28 +0000 (0:00:01.355)       0:55:52.187 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  05:58:30 +0000 (0:00:01.742)       0:55:53.930 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  05:58:31 +0000 (0:00:01.360)       0:55:55.291 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  05:58:34 +0000 (0:00:03.016)       0:55:58.308 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  05:58:36 +0000 (0:00:01.430)       0:55:59.738 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  05:58:37 +0000 (0:00:01.644)       0:56:01.382 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  05:58:38 +0000 (0:00:01.165)       0:56:02.548 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:58:32.222055', 'end': '2026-03-26 05:58:32.279175', 'delta': '0:00:00.057120', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:58:32.799939', 'end': '2026-03-26 05:58:32.846328', 'delta': '0:00:00.046389', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:58:33.349459', 'end': '2026-03-26 05:58:33.398492', 'delta': '0:00:00.049033', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  05:58:40 +0000 (0:00:01.255)       0:56:03.803 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  05:58:41 +0000 (0:00:01.450)       0:56:05.254 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  05:58:42 +0000 (0:00:01.250)       0:56:06.505 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  05:58:44 +0000 (0:00:01.373)       0:56:07.878 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:58:46 +0000 (0:00:02.318)       0:56:10.196 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  05:58:47 +0000 (0:00:01.273)       0:56:11.470 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  05:58:48 +0000 (0:00:01.155)       0:56:12.626 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  05:58:50 +0000 (0:00:01.222)       0:56:13.848 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  05:58:51 +0000 (0:00:01.240)       0:56:15.088 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  05:58:52 +0000 (0:00:01.242)       0:56:16.331 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  05:58:53 +0000 (0:00:01.320)       0:56:17.652 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  05:58:55 +0000 (0:00:01.617)       0:56:19.269 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  05:58:56 +0000 (0:00:01.279)       0:56:20.549 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  05:58:58 +0000 (0:00:01.264)       0:56:21.814 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  05:58:59 +0000 (0:00:01.314)       0:56:23.128 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  05:59:00 +0000 (0:00:01.509)       0:56:24.638 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  05:59:02 +0000 (0:00:01.559)       0:56:26.197 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  05:59:04 +0000 (0:00:01.668)       0:56:27.866 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:59:05 +0000 (0:00:01.240)       0:56:29.107 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:59:07 +0000 (0:00:01.618)       0:56:30.726 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  05:59:08 +0000 (0:00:01.246)       0:56:31.972 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  05:59:09 +0000 (0:00:01.364)       0:56:33.336 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  05:59:10 +0000 (0:00:01.295)       0:56:34.632 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  05:59:13 +0000 (0:00:02.267)       0:56:36.899 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  05:59:14 +0000 (0:00:01.597)       0:56:38.497 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  05:59:16 +0000 (0:00:01.323)       0:56:39.821 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  05:59:17 +0000 (0:00:01.335)       0:56:41.157 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  05:59:18 +0000 (0:00:01.262)       0:56:42.419 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  05:59:19 +0000 (0:00:01.224)       0:56:43.644 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  05:59:21 +0000 (0:00:01.347)       0:56:44.991 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  05:59:23 +0000 (0:00:01.796)       0:56:46.788 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  05:59:24 +0000 (0:00:01.420)       0:56:48.208 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  05:59:25 +0000 (0:00:01.421)       0:56:49.630 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  05:59:27 +0000 (0:00:01.300)       0:56:50.931 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  05:59:29 +0000 (0:00:01.936)       0:56:52.868 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  05:59:31 +0000 (0:00:02.203)       0:56:55.071 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Prevent restarts from the packaging] *************************************\nThursday 26 March 2026  05:59:34 +0000 (0:00:02.707)       0:56:57.778 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  05:59:35 +0000 (0:00:01.214)       0:56:58.993 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  05:59:36 +0000 (0:00:01.540)       0:57:00.533 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  05:59:38 +0000 (0:00:01.267)       0:57:01.800 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  05:59:39 +0000 (0:00:01.300)       0:57:03.101 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  05:59:41 +0000 (0:00:01.581)       0:57:04.682 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  05:59:42 +0000 (0:00:01.705)       0:57:06.387 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  05:59:44 +0000 (0:00:01.631)       0:57:08.019 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  05:59:45 +0000 (0:00:01.243)       0:57:09.262 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  05:59:46 +0000 (0:00:01.247)       0:57:10.510 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  05:59:48 +0000 (0:00:01.277)       0:57:11.787 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  05:59:49 +0000 (0:00:01.612)       0:57:13.400 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  05:59:51 +0000 (0:00:01.638)       0:57:15.039 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  05:59:53 +0000 (0:00:01.646)       0:57:16.685 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  05:59:54 +0000 (0:00:01.256)       0:57:17.942 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  05:59:55 +0000 (0:00:01.324)       0:57:19.266 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  05:59:56 +0000 (0:00:01.264)       0:57:20.531 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  05:59:58 +0000 (0:00:01.272)       0:57:21.804 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  05:59:59 +0000 (0:00:01.682)       0:57:23.486 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  06:00:01 +0000 (0:00:01.238)       0:57:24.725 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  06:00:02 +0000 (0:00:01.668)       0:57:26.395 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  06:00:03 +0000 (0:00:01.243)       0:57:27.638 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  06:00:05 +0000 (0:00:01.346)       0:57:28.985 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  06:00:06 +0000 (0:00:01.278)       0:57:30.263 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  06:00:07 +0000 (0:00:01.298)       0:57:31.562 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  06:00:09 +0000 (0:00:01.164)       0:57:32.726 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  06:00:10 +0000 (0:00:01.284)       0:57:34.011 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  06:00:11 +0000 (0:00:01.170)       0:57:35.181 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  06:00:12 +0000 (0:00:01.203)       0:57:36.385 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  06:00:13 +0000 (0:00:01.167)       0:57:37.552 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  06:00:15 +0000 (0:00:01.195)       0:57:38.748 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  06:00:16 +0000 (0:00:01.229)       0:57:39.978 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  06:00:17 +0000 (0:00:01.673)       0:57:41.651 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  06:00:19 +0000 (0:00:01.238)       0:57:42.890 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  06:00:20 +0000 (0:00:01.227)       0:57:44.118 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  06:00:22 +0000 (0:00:02.049)       0:57:46.167 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  06:00:24 +0000 (0:00:02.379)       0:57:48.546 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  06:00:26 +0000 (0:00:01.473)       0:57:50.020 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  06:00:27 +0000 (0:00:01.306)       0:57:51.326 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  06:00:28 +0000 (0:00:01.248)       0:57:52.575 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  06:00:30 +0000 (0:00:01.880)       0:57:54.455 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  06:00:32 +0000 (0:00:01.591)       0:57:56.047 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  06:00:33 +0000 (0:00:01.242)       0:57:57.289 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  06:00:34 +0000 (0:00:01.351)       0:57:58.641 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  06:00:36 +0000 (0:00:01.187)       0:57:59.829 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  06:00:37 +0000 (0:00:01.249)       0:58:01.078 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  06:00:39 +0000 (0:00:01.920)       0:58:02.999 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  06:00:40 +0000 (0:00:01.223)       0:58:04.222 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  06:00:41 +0000 (0:00:01.258)       0:58:05.480 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  06:00:43 +0000 (0:00:01.271)       0:58:06.752 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  06:00:44 +0000 (0:00:01.239)       0:58:07.992 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  06:00:45 +0000 (0:00:01.323)       0:58:09.315 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  06:00:46 +0000 (0:00:01.289)       0:58:10.605 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  06:00:49 +0000 (0:00:02.705)       0:58:13.310 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  06:00:50 +0000 (0:00:01.215)       0:58:14.526 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  06:00:52 +0000 (0:00:01.462)       0:58:15.988 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  06:00:53 +0000 (0:00:01.268)       0:58:17.257 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  06:00:54 +0000 (0:00:01.286)       0:58:18.543 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  06:00:56 +0000 (0:00:01.272)       0:58:19.815 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  06:00:57 +0000 (0:00:01.256)       0:58:21.072 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  06:00:58 +0000 (0:00:01.263)       0:58:22.336 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  06:01:00 +0000 (0:00:01.352)       0:58:23.688 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  06:01:01 +0000 (0:00:01.283)       0:58:24.972 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  06:01:02 +0000 (0:00:01.310)       0:58:26.283 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  06:01:03 +0000 (0:00:01.254)       0:58:27.537 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  06:01:05 +0000 (0:00:01.226)       0:58:28.763 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  06:01:11 +0000 (0:00:06.635)       0:58:35.399 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  06:01:13 +0000 (0:00:01.343)       0:58:36.742 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  06:01:14 +0000 (0:00:01.652)       0:58:38.394 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  06:01:16 +0000 (0:00:02.002)       0:58:40.397 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  06:01:18 +0000 (0:00:01.311)       0:58:41.708 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  06:01:19 +0000 (0:00:01.260)       0:58:42.969 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  06:01:20 +0000 (0:00:01.598)       0:58:44.567 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  06:01:22 +0000 (0:00:01.212)       0:58:45.779 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  06:01:23 +0000 (0:00:01.208)       0:58:46.987 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  06:01:24 +0000 (0:00:01.235)       0:58:48.222 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  06:01:25 +0000 (0:00:01.290)       0:58:49.513 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  06:01:27 +0000 (0:00:01.316)       0:58:50.830 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  06:01:28 +0000 (0:00:01.315)       0:58:52.146 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  06:01:29 +0000 (0:00:01.304)       0:58:53.450 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  06:01:31 +0000 (0:00:01.304)       0:58:54.755 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  06:01:35 +0000 (0:00:04.640)       0:58:59.396 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  06:01:37 +0000 (0:00:01.357)       0:59:00.753 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  06:01:42 +0000 (0:00:05.214)       0:59:05.968 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  06:01:43 +0000 (0:00:01.242)       0:59:07.211 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:01:45 +0000 (0:00:01.606)       0:59:08.817 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:01:46 +0000 (0:00:01.271)       0:59:10.089 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:01:47 +0000 (0:00:01.305)       0:59:11.395 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:01:49 +0000 (0:00:01.288)       0:59:12.684 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:01:50 +0000 (0:00:01.391)       0:59:14.075 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:01:51 +0000 (0:00:01.420)       0:59:15.495 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:01:53 +0000 (0:00:01.426)       0:59:16.922 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:01:55 +0000 (0:00:01.794)       0:59:18.717 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:01:56 +0000 (0:00:01.338)       0:59:20.055 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  06:01:57 +0000 (0:00:01.506)       0:59:21.562 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nThursday 26 March 2026  06:01:59 +0000 (0:00:01.889)       0:59:23.451 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nThursday 26 March 2026  06:02:01 +0000 (0:00:01.232)       0:59:24.684 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nThursday 26 March 2026  06:02:02 +0000 (0:00:01.436)       0:59:26.121 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:02:04 +0000 (0:00:01.980)       0:59:28.102 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:02:07 +0000 (0:00:03.222)       0:59:31.325 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nThursday 26 March 2026  06:02:09 +0000 (0:00:02.087)       0:59:33.413 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nThursday 26 March 2026  06:02:11 +0000 (0:00:01.583)       0:59:34.996 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nThursday 26 March 2026  06:02:12 +0000 (0:00:01.227)       0:59:36.224 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nThursday 26 March 2026  06:02:14 +0000 (0:00:01.570)       0:59:37.794 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nThursday 26 March 2026  06:02:15 +0000 (0:00:01.442)       0:59:39.237 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nThursday 26 March 2026  06:02:17 +0000 (0:00:02.253)       0:59:41.491 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nThursday 26 March 2026  06:02:19 +0000 (0:00:02.075)       0:59:43.567 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nThursday 26 March 2026  06:02:22 +0000 (0:00:02.371)       0:59:45.939 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nThursday 26 March 2026  06:02:25 +0000 (0:00:03.416)       0:59:49.355 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set max_mds] *************************************************************\nThursday 26 March 2026  06:02:27 +0000 (0:00:01.849)       0:59:51.205 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nPLAY [Upgrade ceph rgws cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  06:02:31 +0000 (0:00:03.502)       0:59:54.708 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  06:02:32 +0000 (0:00:01.142)       0:59:55.850 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  06:02:33 +0000 (0:00:01.452)       0:59:57.302 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:02:34 +0000 (0:00:01.123)       0:59:58.426 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:02:36 +0000 (0:00:01.431)       0:59:59.857 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  06:02:37 +0000 (0:00:01.211)       1:00:01.069 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  06:02:38 +0000 (0:00:01.122)       1:00:02.192 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  06:02:39 +0000 (0:00:01.165)       1:00:03.358 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  06:02:40 +0000 (0:00:01.145)       1:00:04.503 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  06:02:42 +0000 (0:00:01.245)       1:00:05.749 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  06:02:43 +0000 (0:00:01.699)       1:00:07.448 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  06:02:45 +0000 (0:00:01.259)       1:00:08.707 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  06:02:47 +0000 (0:00:02.888)       1:00:11.596 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  06:02:49 +0000 (0:00:01.503)       1:00:13.099 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  06:02:51 +0000 (0:00:01.923)       1:00:15.022 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  06:02:52 +0000 (0:00:01.212)       1:00:16.235 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:02:45.573480', 'end': '2026-03-26 06:02:45.621529', 'delta': '0:00:00.048049', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:02:46.135117', 'end': '2026-03-26 06:02:46.183458', 'delta': '0:00:00.048341', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:02:46.702167', 'end': '2026-03-26 06:02:46.750702', 'delta': '0:00:00.048535', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  06:02:53 +0000 (0:00:01.174)       1:00:17.410 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  06:02:55 +0000 (0:00:01.248)       1:00:18.659 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  06:02:56 +0000 (0:00:01.697)       1:00:20.357 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  06:02:57 +0000 (0:00:01.188)       1:00:21.546 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:02:59 +0000 (0:00:01.989)       1:00:23.535 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  06:03:01 +0000 (0:00:01.183)       1:00:24.718 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  06:03:02 +0000 (0:00:01.161)       1:00:25.880 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:03:03 +0000 (0:00:01.210)       1:00:27.090 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  06:03:04 +0000 (0:00:01.113)       1:00:28.204 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  06:03:05 +0000 (0:00:01.133)       1:00:29.338 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  06:03:06 +0000 (0:00:01.187)       1:00:30.526 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  06:03:07 +0000 (0:00:01.125)       1:00:31.652 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  06:03:09 +0000 (0:00:01.210)       1:00:32.863 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  06:03:10 +0000 (0:00:01.190)       1:00:34.053 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  06:03:11 +0000 (0:00:01.166)       1:00:35.219 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  06:03:13 +0000 (0:00:01.541)       1:00:36.761 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  06:03:14 +0000 (0:00:01.432)       1:00:38.193 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  06:03:16 +0000 (0:00:01.541)       1:00:39.734 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:03:17 +0000 (0:00:01.241)       1:00:40.976 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:03:18 +0000 (0:00:01.543)       1:00:42.520 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:03:20 +0000 (0:00:01.218)       1:00:43.739 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:03:21 +0000 (0:00:01.265)       1:00:45.005 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  06:03:22 +0000 (0:00:01.155)       1:00:46.160 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  06:03:24 +0000 (0:00:02.024)       1:00:48.185 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  06:03:25 +0000 (0:00:01.193)       1:00:49.379 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:03:26 +0000 (0:00:01.136)       1:00:50.516 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:03:27 +0000 (0:00:01.122)       1:00:51.638 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:03:29 +0000 (0:00:01.242)       1:00:52.881 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:03:30 +0000 (0:00:01.155)       1:00:54.037 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:03:31 +0000 (0:00:01.210)       1:00:55.247 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:03:33 +0000 (0:00:01.437)       1:00:56.685 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:03:34 +0000 (0:00:01.470)       1:00:58.155 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:03:35 +0000 (0:00:01.422)       1:00:59.578 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:03:37 +0000 (0:00:01.176)       1:01:00.755 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  06:03:38 +0000 (0:00:01.330)       1:01:02.085 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  06:03:40 +0000 (0:00:02.193)       1:01:04.278 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Stop ceph rgw when upgrading from stable-3.2] ****************************\nThursday 26 March 2026  06:03:43 +0000 (0:00:02.646)       1:01:06.924 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Stop ceph rgw (pt. 1)] ***************************************************\nThursday 26 March 2026  06:03:45 +0000 (0:00:02.315)       1:01:09.240 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [Stop ceph rgw (pt. 2)] ***************************************************\nThursday 26 March 2026  06:03:48 +0000 (0:00:02.873)       1:01:12.113 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  06:03:50 +0000 (0:00:02.383)       1:01:14.497 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  06:03:51 +0000 (0:00:01.109)       1:01:15.606 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  06:03:53 +0000 (0:00:01.146)       1:01:16.753 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  06:03:54 +0000 (0:00:01.121)       1:01:17.875 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  06:03:55 +0000 (0:00:01.497)       1:01:19.372 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  06:03:57 +0000 (0:00:01.557)       1:01:20.930 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  06:03:58 +0000 (0:00:01.501)       1:01:22.431 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  06:03:59 +0000 (0:00:01.125)       1:01:23.556 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  06:04:01 +0000 (0:00:01.125)       1:01:24.682 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  06:04:02 +0000 (0:00:01.108)       1:01:25.790 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  06:04:03 +0000 (0:00:01.514)       1:01:27.305 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  06:04:05 +0000 (0:00:01.503)       1:01:28.809 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  06:04:06 +0000 (0:00:01.183)       1:01:29.993 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  06:04:07 +0000 (0:00:01.100)       1:01:31.093 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  06:04:08 +0000 (0:00:01.127)       1:01:32.220 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  06:04:09 +0000 (0:00:01.126)       1:01:33.347 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  06:04:10 +0000 (0:00:01.224)       1:01:34.571 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  06:04:12 +0000 (0:00:01.119)       1:01:35.691 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  06:04:13 +0000 (0:00:01.142)       1:01:36.833 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  06:04:14 +0000 (0:00:01.139)       1:01:37.973 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  06:04:15 +0000 (0:00:01.178)       1:01:39.151 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  06:04:16 +0000 (0:00:01.186)       1:01:40.338 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  06:04:17 +0000 (0:00:01.148)       1:01:41.487 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  06:04:18 +0000 (0:00:01.114)       1:01:42.602 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  06:04:20 +0000 (0:00:01.242)       1:01:43.844 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  06:04:21 +0000 (0:00:01.216)       1:01:45.061 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  06:04:22 +0000 (0:00:01.141)       1:01:46.202 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  06:04:23 +0000 (0:00:01.129)       1:01:47.331 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  06:04:24 +0000 (0:00:01.112)       1:01:48.443 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  06:04:25 +0000 (0:00:01.131)       1:01:49.575 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  06:04:27 +0000 (0:00:01.103)       1:01:50.679 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  06:04:28 +0000 (0:00:01.125)       1:01:51.804 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  06:04:29 +0000 (0:00:01.126)       1:01:52.931 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  06:04:30 +0000 (0:00:01.168)       1:01:54.100 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  06:04:32 +0000 (0:00:01.900)       1:01:56.000 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  06:04:34 +0000 (0:00:02.277)       1:01:58.278 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  06:04:35 +0000 (0:00:01.243)       1:01:59.521 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  06:04:36 +0000 (0:00:01.132)       1:02:00.654 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  06:04:38 +0000 (0:00:01.128)       1:02:01.783 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  06:04:39 +0000 (0:00:01.787)       1:02:03.571 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  06:04:41 +0000 (0:00:01.497)       1:02:05.068 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  06:04:42 +0000 (0:00:01.151)       1:02:06.219 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  06:04:43 +0000 (0:00:01.131)       1:02:07.351 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  06:04:44 +0000 (0:00:01.149)       1:02:08.500 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  06:04:45 +0000 (0:00:01.145)       1:02:09.645 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  06:04:47 +0000 (0:00:01.938)       1:02:11.584 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  06:04:49 +0000 (0:00:01.231)       1:02:12.816 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  06:04:50 +0000 (0:00:01.358)       1:02:14.175 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  06:04:51 +0000 (0:00:01.202)       1:02:15.377 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  06:04:52 +0000 (0:00:01.198)       1:02:16.576 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  06:04:54 +0000 (0:00:01.230)       1:02:17.807 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  06:04:55 +0000 (0:00:01.201)       1:02:19.008 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  06:04:57 +0000 (0:00:02.504)       1:02:21.513 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  06:04:59 +0000 (0:00:01.152)       1:02:22.665 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  06:05:00 +0000 (0:00:01.113)       1:02:23.779 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  06:05:01 +0000 (0:00:01.182)       1:02:24.962 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  06:05:02 +0000 (0:00:01.161)       1:02:26.124 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  06:05:03 +0000 (0:00:01.185)       1:02:27.310 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  06:05:04 +0000 (0:00:01.160)       1:02:28.470 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  06:05:05 +0000 (0:00:01.153)       1:02:29.624 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  06:05:07 +0000 (0:00:01.262)       1:02:30.886 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  06:05:08 +0000 (0:00:01.157)       1:02:32.043 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  06:05:09 +0000 (0:00:01.163)       1:02:33.207 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  06:05:10 +0000 (0:00:01.177)       1:02:34.385 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  06:05:11 +0000 (0:00:01.115)       1:02:35.500 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  06:05:18 +0000 (0:00:06.698)       1:02:42.199 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  06:05:19 +0000 (0:00:01.155)       1:02:43.354 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  06:05:21 +0000 (0:00:01.521)       1:02:44.875 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  06:05:23 +0000 (0:00:01.949)       1:02:46.824 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  06:05:24 +0000 (0:00:01.158)       1:02:47.983 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  06:05:25 +0000 (0:00:01.278)       1:02:49.261 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  06:05:26 +0000 (0:00:01.123)       1:02:50.385 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  06:05:27 +0000 (0:00:01.139)       1:02:51.525 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  06:05:28 +0000 (0:00:01.131)       1:02:52.657 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  06:05:30 +0000 (0:00:01.101)       1:02:53.758 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  06:05:31 +0000 (0:00:01.206)       1:02:54.965 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  06:05:32 +0000 (0:00:01.150)       1:02:56.115 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  06:05:33 +0000 (0:00:01.120)       1:02:57.236 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  06:05:34 +0000 (0:00:01.133)       1:02:58.369 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  06:05:35 +0000 (0:00:01.172)       1:02:59.542 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  06:05:40 +0000 (0:00:04.512)       1:03:04.055 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  06:05:41 +0000 (0:00:01.148)       1:03:05.204 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  06:05:46 +0000 (0:00:05.259)       1:03:10.463 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  06:05:48 +0000 (0:00:01.309)       1:03:11.773 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:05:49 +0000 (0:00:01.196)       1:03:12.970 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:05:50 +0000 (0:00:01.248)       1:03:14.218 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:05:51 +0000 (0:00:01.127)       1:03:15.346 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:05:52 +0000 (0:00:01.210)       1:03:16.557 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:05:54 +0000 (0:00:01.257)       1:03:17.814 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:05:55 +0000 (0:00:01.473)       1:03:19.288 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:05:57 +0000 (0:00:01.405)       1:03:20.694 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:05:58 +0000 (0:00:01.410)       1:03:22.104 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:05:59 +0000 (0:00:01.170)       1:03:23.275 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  06:06:00 +0000 (0:00:01.369)       1:03:24.645 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nThursday 26 March 2026  06:06:02 +0000 (0:00:01.946)       1:03:26.591 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:06:04 +0000 (0:00:01.472)       1:03:28.064 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:06:07 +0000 (0:00:03.236)       1:03:31.300 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nThursday 26 March 2026  06:06:09 +0000 (0:00:02.013)       1:03:33.314 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nThursday 26 March 2026  06:06:10 +0000 (0:00:01.169)       1:03:34.484 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nThursday 26 March 2026  06:06:12 +0000 (0:00:01.484)       1:03:35.968 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nThursday 26 March 2026  06:06:13 +0000 (0:00:01.599)       1:03:37.568 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:06:19 +0000 (0:00:05.222)       1:03:42.790 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:06:22 +0000 (0:00:03.171)       1:03:45.961 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nThursday 26 March 2026  06:06:24 +0000 (0:00:02.040)       1:03:48.002 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nThursday 26 March 2026  06:06:25 +0000 (0:00:01.630)       1:03:49.633 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nThursday 26 March 2026  06:06:27 +0000 (0:00:01.612)       1:03:51.245 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nThursday 26 March 2026  06:06:29 +0000 (0:00:01.666)       1:03:52.912 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nThursday 26 March 2026  06:07:00 +0000 (0:00:31.259)       1:04:24.172 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nThursday 26 March 2026  06:07:01 +0000 (0:00:01.146)       1:04:25.319 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nThursday 26 March 2026  06:07:02 +0000 (0:00:01.146)       1:04:26.465 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nThursday 26 March 2026  06:07:04 +0000 (0:00:01.491)       1:04:27.957 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nThursday 26 March 2026  06:07:05 +0000 (0:00:01.521)       1:04:29.478 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nThursday 26 March 2026  06:07:07 +0000 (0:00:02.099)       1:04:31.577 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nThursday 26 March 2026  06:07:09 +0000 (0:00:02.027)       1:04:33.605 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nThursday 26 March 2026  06:07:12 +0000 (0:00:02.309)       1:04:35.914 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\nPLAY [Upgrade ceph rgws cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  06:07:15 +0000 (0:00:03.155)       1:04:39.070 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  06:07:16 +0000 (0:00:01.119)       1:04:40.189 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  06:07:17 +0000 (0:00:01.429)       1:04:41.619 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:07:19 +0000 (0:00:01.150)       1:04:42.770 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:07:20 +0000 (0:00:01.465)       1:04:44.236 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  06:07:21 +0000 (0:00:01.143)       1:04:45.379 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  06:07:22 +0000 (0:00:01.125)       1:04:46.505 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  06:07:24 +0000 (0:00:01.207)       1:04:47.713 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  06:07:25 +0000 (0:00:01.143)       1:04:48.857 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  06:07:26 +0000 (0:00:01.132)       1:04:49.990 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  06:07:28 +0000 (0:00:02.144)       1:04:52.134 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  06:07:29 +0000 (0:00:01.245)       1:04:53.379 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  06:07:32 +0000 (0:00:02.918)       1:04:56.298 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  06:07:34 +0000 (0:00:01.492)       1:04:57.791 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  06:07:35 +0000 (0:00:01.734)       1:04:59.526 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  06:07:37 +0000 (0:00:01.227)       1:05:00.753 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:07:30.238507', 'end': '2026-03-26 06:07:30.291933', 'delta': '0:00:00.053426', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:07:30.798086', 'end': '2026-03-26 06:07:30.854639', 'delta': '0:00:00.056553', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:07:31.399026', 'end': '2026-03-26 06:07:31.431938', 'delta': '0:00:00.032912', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  06:07:38 +0000 (0:00:01.202)       1:05:01.955 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  06:07:39 +0000 (0:00:01.272)       1:05:03.228 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  06:07:40 +0000 (0:00:01.269)       1:05:04.497 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  06:07:41 +0000 (0:00:01.151)       1:05:05.648 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:07:43 +0000 (0:00:01.996)       1:05:07.645 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  06:07:45 +0000 (0:00:01.162)       1:05:08.808 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  06:07:46 +0000 (0:00:01.121)       1:05:09.929 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:07:48 +0000 (0:00:01.740)       1:05:11.669 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  06:07:49 +0000 (0:00:01.177)       1:05:12.847 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  06:07:50 +0000 (0:00:01.128)       1:05:13.975 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  06:07:51 +0000 (0:00:01.151)       1:05:15.127 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  06:07:52 +0000 (0:00:01.147)       1:05:16.274 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  06:07:53 +0000 (0:00:01.172)       1:05:17.447 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  06:07:55 +0000 (0:00:01.213)       1:05:18.661 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  06:07:56 +0000 (0:00:01.163)       1:05:19.824 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  06:07:57 +0000 (0:00:01.331)       1:05:21.155 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  06:07:58 +0000 (0:00:01.444)       1:05:22.600 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  06:08:00 +0000 (0:00:01.517)       1:05:24.119 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:08:01 +0000 (0:00:01.150)       1:05:25.270 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:08:03 +0000 (0:00:01.472)       1:05:26.743 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:08:04 +0000 (0:00:01.310)       1:05:28.054 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:08:05 +0000 (0:00:01.315)       1:05:29.369 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  06:08:06 +0000 (0:00:01.125)       1:05:30.495 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  06:08:08 +0000 (0:00:01.668)       1:05:32.164 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  06:08:10 +0000 (0:00:01.664)       1:05:33.829 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:08:11 +0000 (0:00:01.166)       1:05:34.996 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:08:12 +0000 (0:00:01.143)       1:05:36.139 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:08:13 +0000 (0:00:01.134)       1:05:37.274 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:08:14 +0000 (0:00:01.136)       1:05:38.410 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:08:16 +0000 (0:00:01.297)       1:05:39.708 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:08:17 +0000 (0:00:01.393)       1:05:41.101 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:08:18 +0000 (0:00:01.517)       1:05:42.619 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:08:20 +0000 (0:00:01.362)       1:05:43.982 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:08:21 +0000 (0:00:00.921)       1:05:44.903 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  06:08:22 +0000 (0:00:01.229)       1:05:46.133 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  06:08:24 +0000 (0:00:01.797)       1:05:47.930 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Stop ceph rgw when upgrading from stable-3.2] ****************************\nThursday 26 March 2026  06:08:26 +0000 (0:00:02.534)       1:05:50.464 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Stop ceph rgw (pt. 1)] ***************************************************\nThursday 26 March 2026  06:08:28 +0000 (0:00:01.943)       1:05:52.408 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [Stop ceph rgw (pt. 2)] ***************************************************\nThursday 26 March 2026  06:08:32 +0000 (0:00:03.633)       1:05:56.041 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  06:08:34 +0000 (0:00:01.931)       1:05:57.973 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  06:08:35 +0000 (0:00:01.119)       1:05:59.093 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  06:08:36 +0000 (0:00:01.099)       1:06:00.192 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  06:08:37 +0000 (0:00:01.128)       1:06:01.321 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  06:08:39 +0000 (0:00:01.588)       1:06:02.909 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  06:08:40 +0000 (0:00:01.522)       1:06:04.431 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  06:08:42 +0000 (0:00:01.521)       1:06:05.953 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  06:08:43 +0000 (0:00:01.161)       1:06:07.114 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  06:08:44 +0000 (0:00:01.145)       1:06:08.260 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  06:08:45 +0000 (0:00:01.167)       1:06:09.427 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  06:08:47 +0000 (0:00:01.557)       1:06:10.985 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  06:08:48 +0000 (0:00:01.513)       1:06:12.499 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  06:08:49 +0000 (0:00:00.778)       1:06:13.277 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  06:08:50 +0000 (0:00:00.806)       1:06:14.083 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  06:08:51 +0000 (0:00:00.838)       1:06:14.922 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  06:08:52 +0000 (0:00:00.832)       1:06:15.754 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  06:08:52 +0000 (0:00:00.845)       1:06:16.600 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  06:08:53 +0000 (0:00:00.846)       1:06:17.446 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  06:08:54 +0000 (0:00:00.815)       1:06:18.262 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  06:08:55 +0000 (0:00:00.773)       1:06:19.035 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  06:08:56 +0000 (0:00:00.814)       1:06:19.850 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  06:08:56 +0000 (0:00:00.805)       1:06:20.656 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  06:08:57 +0000 (0:00:00.758)       1:06:21.415 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  06:08:58 +0000 (0:00:00.853)       1:06:22.268 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  06:08:59 +0000 (0:00:00.786)       1:06:23.054 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  06:09:00 +0000 (0:00:00.765)       1:06:23.820 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  06:09:00 +0000 (0:00:00.764)       1:06:24.585 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  06:09:01 +0000 (0:00:00.809)       1:06:25.395 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  06:09:02 +0000 (0:00:00.805)       1:06:26.200 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  06:09:03 +0000 (0:00:00.864)       1:06:27.065 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  06:09:04 +0000 (0:00:00.840)       1:06:27.905 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  06:09:05 +0000 (0:00:00.811)       1:06:28.717 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  06:09:05 +0000 (0:00:00.803)       1:06:29.520 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  06:09:06 +0000 (0:00:00.774)       1:06:30.295 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  06:09:08 +0000 (0:00:01.596)       1:06:31.892 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  06:09:10 +0000 (0:00:01.890)       1:06:33.782 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  06:09:11 +0000 (0:00:01.174)       1:06:34.956 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  06:09:12 +0000 (0:00:01.164)       1:06:36.121 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  06:09:13 +0000 (0:00:01.144)       1:06:37.265 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  06:09:15 +0000 (0:00:01.872)       1:06:39.138 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  06:09:16 +0000 (0:00:01.439)       1:06:40.578 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  06:09:18 +0000 (0:00:01.715)       1:06:42.294 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  06:09:19 +0000 (0:00:00.823)       1:06:43.117 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  06:09:20 +0000 (0:00:00.793)       1:06:43.910 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  06:09:21 +0000 (0:00:01.138)       1:06:45.049 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  06:09:23 +0000 (0:00:01.847)       1:06:46.897 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  06:09:24 +0000 (0:00:01.138)       1:06:48.035 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  06:09:25 +0000 (0:00:01.135)       1:06:49.171 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  06:09:26 +0000 (0:00:01.141)       1:06:50.313 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  06:09:27 +0000 (0:00:01.187)       1:06:51.501 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  06:09:29 +0000 (0:00:01.179)       1:06:52.681 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  06:09:29 +0000 (0:00:00.807)       1:06:53.488 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  06:09:31 +0000 (0:00:02.130)       1:06:55.619 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  06:09:32 +0000 (0:00:00.856)       1:06:56.475 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  06:09:33 +0000 (0:00:01.162)       1:06:57.637 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  06:09:35 +0000 (0:00:01.131)       1:06:58.769 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  06:09:36 +0000 (0:00:01.155)       1:06:59.924 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  06:09:37 +0000 (0:00:01.174)       1:07:01.099 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  06:09:38 +0000 (0:00:01.181)       1:07:02.281 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  06:09:39 +0000 (0:00:01.119)       1:07:03.400 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  06:09:40 +0000 (0:00:01.223)       1:07:04.624 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  06:09:42 +0000 (0:00:01.161)       1:07:05.786 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  06:09:43 +0000 (0:00:01.113)       1:07:06.900 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  06:09:44 +0000 (0:00:00.829)       1:07:07.730 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  06:09:45 +0000 (0:00:01.306)       1:07:09.036 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  06:09:51 +0000 (0:00:06.102)       1:07:15.139 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  06:09:52 +0000 (0:00:01.117)       1:07:16.256 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  06:09:54 +0000 (0:00:01.529)       1:07:17.786 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  06:09:55 +0000 (0:00:01.707)       1:07:19.494 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  06:09:56 +0000 (0:00:00.819)       1:07:20.314 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  06:09:57 +0000 (0:00:00.766)       1:07:21.080 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  06:09:58 +0000 (0:00:00.771)       1:07:21.851 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  06:09:59 +0000 (0:00:00.806)       1:07:22.658 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  06:09:59 +0000 (0:00:00.803)       1:07:23.461 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  06:10:00 +0000 (0:00:00.810)       1:07:24.272 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  06:10:01 +0000 (0:00:00.829)       1:07:25.102 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  06:10:02 +0000 (0:00:00.815)       1:07:25.917 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  06:10:03 +0000 (0:00:00.764)       1:07:26.682 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  06:10:03 +0000 (0:00:00.806)       1:07:27.490 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  06:10:04 +0000 (0:00:00.810)       1:07:28.300 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  06:10:08 +0000 (0:00:04.119)       1:07:32.420 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  06:10:09 +0000 (0:00:00.795)       1:07:33.216 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  06:10:13 +0000 (0:00:04.429)       1:07:37.645 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  06:10:14 +0000 (0:00:00.807)       1:07:38.453 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:10:15 +0000 (0:00:00.793)       1:07:39.246 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:10:16 +0000 (0:00:00.817)       1:07:40.064 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:10:17 +0000 (0:00:00.815)       1:07:40.880 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:10:18 +0000 (0:00:00.821)       1:07:41.701 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:10:19 +0000 (0:00:01.058)       1:07:42.759 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:10:20 +0000 (0:00:01.076)       1:07:43.836 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:10:21 +0000 (0:00:01.131)       1:07:44.967 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:10:22 +0000 (0:00:01.087)       1:07:46.055 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:10:23 +0000 (0:00:00.843)       1:07:46.899 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  06:10:24 +0000 (0:00:01.042)       1:07:47.942 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nThursday 26 March 2026  06:10:25 +0000 (0:00:01.435)       1:07:49.378 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:10:26 +0000 (0:00:01.141)       1:07:50.519 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:10:30 +0000 (0:00:03.219)       1:07:53.739 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nThursday 26 March 2026  06:10:32 +0000 (0:00:01.950)       1:07:55.689 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nThursday 26 March 2026  06:10:32 +0000 (0:00:00.749)       1:07:56.438 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nThursday 26 March 2026  06:10:34 +0000 (0:00:01.315)       1:07:57.754 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nThursday 26 March 2026  06:10:35 +0000 (0:00:01.590)       1:07:59.345 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:10:40 +0000 (0:00:05.213)       1:08:04.559 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:10:43 +0000 (0:00:03.044)       1:08:07.604 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nThursday 26 March 2026  06:10:45 +0000 (0:00:01.639)       1:08:09.244 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nThursday 26 March 2026  06:10:46 +0000 (0:00:01.160)       1:08:10.404 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nThursday 26 March 2026  06:10:48 +0000 (0:00:01.647)       1:08:12.052 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nThursday 26 March 2026  06:10:50 +0000 (0:00:01.995)       1:08:14.047 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nThursday 26 March 2026  06:11:21 +0000 (0:00:31.450)       1:08:45.498 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nThursday 26 March 2026  06:11:22 +0000 (0:00:00.762)       1:08:46.260 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nThursday 26 March 2026  06:11:23 +0000 (0:00:00.772)       1:08:47.032 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nThursday 26 March 2026  06:11:24 +0000 (0:00:01.262)       1:08:48.295 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-4\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nThursday 26 March 2026  06:11:25 +0000 (0:00:01.163)       1:08:49.458 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nThursday 26 March 2026  06:11:27 +0000 (0:00:02.068)       1:08:51.526 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nThursday 26 March 2026  06:11:29 +0000 (0:00:01.867)       1:08:53.394 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nThursday 26 March 2026  06:11:31 +0000 (0:00:02.209)       1:08:55.603 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nPLAY [Upgrade ceph rgws cluster] ***********************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  06:11:34 +0000 (0:00:02.821)       1:08:58.425 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nThursday 26 March 2026  06:11:35 +0000 (0:00:01.179)       1:08:59.604 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nThursday 26 March 2026  06:11:37 +0000 (0:00:01.538)       1:09:01.143 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:11:38 +0000 (0:00:01.152)       1:09:02.295 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:11:40 +0000 (0:00:01.490)       1:09:03.786 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nThursday 26 March 2026  06:11:41 +0000 (0:00:01.188)       1:09:04.974 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nThursday 26 March 2026  06:11:42 +0000 (0:00:01.179)       1:09:06.154 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nThursday 26 March 2026  06:11:43 +0000 (0:00:01.194)       1:09:07.348 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nThursday 26 March 2026  06:11:44 +0000 (0:00:01.137)       1:09:08.486 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nThursday 26 March 2026  06:11:45 +0000 (0:00:01.143)       1:09:09.629 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nThursday 26 March 2026  06:11:47 +0000 (0:00:01.764)       1:09:11.394 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nThursday 26 March 2026  06:11:49 +0000 (0:00:01.358)       1:09:12.752 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nThursday 26 March 2026  06:11:52 +0000 (0:00:03.297)       1:09:16.050 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nThursday 26 March 2026  06:11:53 +0000 (0:00:01.406)       1:09:17.456 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nThursday 26 March 2026  06:11:55 +0000 (0:00:02.092)       1:09:19.548 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nThursday 26 March 2026  06:11:57 +0000 (0:00:01.154)       1:09:20.703 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:11:49.624637', 'end': '2026-03-26 06:11:49.669646', 'delta': '0:00:00.045009', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:11:50.174575', 'end': '2026-03-26 06:11:50.235008', 'delta': '0:00:00.060433', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:11:51.182319', 'end': '2026-03-26 06:11:51.231785', 'delta': '0:00:00.049466', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nThursday 26 March 2026  06:11:58 +0000 (0:00:01.315)       1:09:22.018 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nThursday 26 March 2026  06:11:59 +0000 (0:00:01.280)       1:09:23.299 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nThursday 26 March 2026  06:12:00 +0000 (0:00:01.259)       1:09:24.558 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nThursday 26 March 2026  06:12:02 +0000 (0:00:01.133)       1:09:25.692 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:12:04 +0000 (0:00:02.020)       1:09:27.712 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nThursday 26 March 2026  06:12:05 +0000 (0:00:01.150)       1:09:28.863 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nThursday 26 March 2026  06:12:06 +0000 (0:00:01.123)       1:09:29.987 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nThursday 26 March 2026  06:12:07 +0000 (0:00:01.236)       1:09:31.223 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nThursday 26 March 2026  06:12:08 +0000 (0:00:01.181)       1:09:32.405 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nThursday 26 March 2026  06:12:09 +0000 (0:00:01.121)       1:09:33.526 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nThursday 26 March 2026  06:12:11 +0000 (0:00:01.182)       1:09:34.709 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nThursday 26 March 2026  06:12:12 +0000 (0:00:01.137)       1:09:35.846 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nThursday 26 March 2026  06:12:13 +0000 (0:00:01.163)       1:09:37.009 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nThursday 26 March 2026  06:12:14 +0000 (0:00:01.302)       1:09:38.312 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nThursday 26 March 2026  06:12:15 +0000 (0:00:01.202)       1:09:39.514 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nThursday 26 March 2026  06:12:17 +0000 (0:00:01.407)       1:09:40.922 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nThursday 26 March 2026  06:12:18 +0000 (0:00:01.383)       1:09:42.306 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nThursday 26 March 2026  06:12:20 +0000 (0:00:01.531)       1:09:43.838 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:12:21 +0000 (0:00:01.150)       1:09:44.988 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:12:22 +0000 (0:00:01.511)       1:09:46.499 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nThursday 26 March 2026  06:12:23 +0000 (0:00:01.109)       1:09:47.609 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nThursday 26 March 2026  06:12:25 +0000 (0:00:01.263)       1:09:48.873 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nThursday 26 March 2026  06:12:26 +0000 (0:00:01.189)       1:09:50.062 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nThursday 26 March 2026  06:12:28 +0000 (0:00:02.086)       1:09:52.149 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nThursday 26 March 2026  06:12:29 +0000 (0:00:01.182)       1:09:53.331 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:12:30 +0000 (0:00:01.303)       1:09:54.635 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:12:32 +0000 (0:00:01.222)       1:09:55.858 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:12:33 +0000 (0:00:01.130)       1:09:56.988 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:12:34 +0000 (0:00:01.151)       1:09:58.140 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:12:35 +0000 (0:00:01.247)       1:09:59.388 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:12:37 +0000 (0:00:01.441)       1:10:00.830 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:12:38 +0000 (0:00:01.414)       1:10:02.245 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:12:40 +0000 (0:00:01.423)       1:10:03.669 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:12:41 +0000 (0:00:01.177)       1:10:04.846 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nThursday 26 March 2026  06:12:42 +0000 (0:00:01.339)       1:10:06.186 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nThursday 26 March 2026  06:12:44 +0000 (0:00:02.218)       1:10:08.405 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Stop ceph rgw when upgrading from stable-3.2] ****************************\nThursday 26 March 2026  06:12:47 +0000 (0:00:02.296)       1:10:10.701 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Stop ceph rgw (pt. 1)] ***************************************************\nThursday 26 March 2026  06:12:49 +0000 (0:00:02.048)       1:10:12.750 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [Stop ceph rgw (pt. 2)] ***************************************************\nThursday 26 March 2026  06:12:51 +0000 (0:00:02.507)       1:10:15.258 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  06:12:53 +0000 (0:00:01.940)       1:10:17.198 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  06:12:54 +0000 (0:00:01.092)       1:10:18.291 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  06:12:55 +0000 (0:00:01.144)       1:10:19.435 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  06:12:56 +0000 (0:00:01.122)       1:10:20.558 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  06:12:58 +0000 (0:00:01.559)       1:10:22.118 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  06:12:59 +0000 (0:00:01.535)       1:10:23.653 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  06:13:01 +0000 (0:00:01.510)       1:10:25.164 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  06:13:02 +0000 (0:00:01.141)       1:10:26.306 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  06:13:03 +0000 (0:00:01.111)       1:10:27.417 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  06:13:04 +0000 (0:00:01.175)       1:10:28.592 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  06:13:06 +0000 (0:00:01.522)       1:10:30.114 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  06:13:07 +0000 (0:00:01.497)       1:10:31.612 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  06:13:08 +0000 (0:00:00.751)       1:10:32.363 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  06:13:09 +0000 (0:00:00.756)       1:10:33.120 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  06:13:10 +0000 (0:00:00.820)       1:10:33.941 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  06:13:11 +0000 (0:00:00.831)       1:10:34.772 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  06:13:11 +0000 (0:00:00.808)       1:10:35.581 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  06:13:12 +0000 (0:00:00.800)       1:10:36.382 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  06:13:13 +0000 (0:00:00.804)       1:10:37.186 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  06:13:14 +0000 (0:00:00.752)       1:10:37.939 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  06:13:15 +0000 (0:00:00.764)       1:10:38.704 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_repository.yml] **************************\nThursday 26 March 2026  06:13:16 +0000 (0:00:00.975)       1:10:39.679 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_redhat_packages.yml] **************\nThursday 26 March 2026  06:13:16 +0000 (0:00:00.792)       1:10:40.471 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_suse_packages.yml] ****************\nThursday 26 March 2026  06:13:17 +0000 (0:00:00.804)       1:10:41.275 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include installs/install_on_debian.yml] ********************\nThursday 26 March 2026  06:13:18 +0000 (0:00:00.793)       1:10:42.069 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************\nThursday 26 March 2026  06:13:19 +0000 (0:00:00.810)       1:10:42.880 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Get ceph version] ******************************************\nThursday 26 March 2026  06:13:20 +0000 (0:00:00.781)       1:10:43.661 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_version] *************************************\nThursday 26 March 2026  06:13:20 +0000 (0:00:00.811)       1:10:44.473 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***\nThursday 26 March 2026  06:13:21 +0000 (0:00:00.763)       1:10:45.236 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include create_rbd_client_dir.yml] *************************\nThursday 26 March 2026  06:13:22 +0000 (0:00:00.778)       1:10:46.015 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_cluster_name.yml] ************************\nThursday 26 March 2026  06:13:23 +0000 (0:00:00.780)       1:10:46.796 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include configure_memory_allocator.yml] ********************\nThursday 26 March 2026  06:13:23 +0000 (0:00:00.780)       1:10:47.576 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-common : Include selinux.yml] ***************************************\nThursday 26 March 2026  06:13:24 +0000 (0:00:00.779)       1:10:48.355 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nThursday 26 March 2026  06:13:25 +0000 (0:00:00.797)       1:10:49.153 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nThursday 26 March 2026  06:13:27 +0000 (0:00:01.669)       1:10:50.822 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nThursday 26 March 2026  06:13:29 +0000 (0:00:01.902)       1:10:52.724 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nThursday 26 March 2026  06:13:30 +0000 (0:00:01.127)       1:10:53.852 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nThursday 26 March 2026  06:13:31 +0000 (0:00:01.227)       1:10:55.079 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nThursday 26 March 2026  06:13:32 +0000 (0:00:01.140)       1:10:56.220 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nThursday 26 March 2026  06:13:34 +0000 (0:00:01.774)       1:10:57.995 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nThursday 26 March 2026  06:13:35 +0000 (0:00:01.489)       1:10:59.484 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nThursday 26 March 2026  06:13:36 +0000 (0:00:01.121)       1:11:00.605 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nThursday 26 March 2026  06:13:37 +0000 (0:00:00.823)       1:11:01.429 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nThursday 26 March 2026  06:13:38 +0000 (0:00:00.769)       1:11:02.198 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nThursday 26 March 2026  06:13:39 +0000 (0:00:01.153)       1:11:03.352 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nThursday 26 March 2026  06:13:41 +0000 (0:00:01.901)       1:11:05.254 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nThursday 26 March 2026  06:13:42 +0000 (0:00:01.154)       1:11:06.409 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nThursday 26 March 2026  06:13:43 +0000 (0:00:01.154)       1:11:07.563 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nThursday 26 March 2026  06:13:45 +0000 (0:00:01.164)       1:11:08.727 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nThursday 26 March 2026  06:13:46 +0000 (0:00:01.225)       1:11:09.953 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nThursday 26 March 2026  06:13:47 +0000 (0:00:01.149)       1:11:11.103 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nThursday 26 March 2026  06:13:48 +0000 (0:00:00.786)       1:11:11.889 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nThursday 26 March 2026  06:13:50 +0000 (0:00:02.128)       1:11:14.018 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nThursday 26 March 2026  06:13:51 +0000 (0:00:00.789)       1:11:14.807 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nThursday 26 March 2026  06:13:52 +0000 (0:00:01.153)       1:11:15.960 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nThursday 26 March 2026  06:13:53 +0000 (0:00:01.139)       1:11:17.100 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nThursday 26 March 2026  06:13:54 +0000 (0:00:01.133)       1:11:18.233 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nThursday 26 March 2026  06:13:55 +0000 (0:00:01.304)       1:11:19.538 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nThursday 26 March 2026  06:13:57 +0000 (0:00:01.134)       1:11:20.673 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nThursday 26 March 2026  06:13:58 +0000 (0:00:01.132)       1:11:21.805 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nThursday 26 March 2026  06:13:59 +0000 (0:00:01.218)       1:11:23.023 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nThursday 26 March 2026  06:14:00 +0000 (0:00:01.210)       1:11:24.234 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nThursday 26 March 2026  06:14:01 +0000 (0:00:01.177)       1:11:25.412 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nThursday 26 March 2026  06:14:02 +0000 (0:00:00.832)       1:11:26.244 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nThursday 26 March 2026  06:14:03 +0000 (0:00:01.172)       1:11:27.416 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nThursday 26 March 2026  06:14:09 +0000 (0:00:06.082)       1:11:33.499 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nThursday 26 March 2026  06:14:10 +0000 (0:00:01.095)       1:11:34.594 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nThursday 26 March 2026  06:14:12 +0000 (0:00:01.634)       1:11:36.229 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nThursday 26 March 2026  06:14:14 +0000 (0:00:01.585)       1:11:37.815 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nThursday 26 March 2026  06:14:14 +0000 (0:00:00.830)       1:11:38.645 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nThursday 26 March 2026  06:14:15 +0000 (0:00:00.800)       1:11:39.446 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nThursday 26 March 2026  06:14:16 +0000 (0:00:00.870)       1:11:40.317 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nThursday 26 March 2026  06:14:17 +0000 (0:00:00.774)       1:11:41.091 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nThursday 26 March 2026  06:14:18 +0000 (0:00:00.809)       1:11:41.901 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nThursday 26 March 2026  06:14:19 +0000 (0:00:00.775)       1:11:42.677 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nThursday 26 March 2026  06:14:19 +0000 (0:00:00.779)       1:11:43.457 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nThursday 26 March 2026  06:14:20 +0000 (0:00:00.818)       1:11:44.275 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nThursday 26 March 2026  06:14:21 +0000 (0:00:00.830)       1:11:45.105 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nThursday 26 March 2026  06:14:22 +0000 (0:00:00.794)       1:11:45.900 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nThursday 26 March 2026  06:14:23 +0000 (0:00:00.788)       1:11:46.689 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nThursday 26 March 2026  06:14:27 +0000 (0:00:04.177)       1:11:50.866 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nThursday 26 March 2026  06:14:28 +0000 (0:00:00.932)       1:11:51.798 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nThursday 26 March 2026  06:14:32 +0000 (0:00:04.731)       1:11:56.530 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nThursday 26 March 2026  06:14:33 +0000 (0:00:00.778)       1:11:57.308 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nThursday 26 March 2026  06:14:34 +0000 (0:00:00.838)       1:11:58.147 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nThursday 26 March 2026  06:14:35 +0000 (0:00:00.809)       1:11:58.957 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nThursday 26 March 2026  06:14:36 +0000 (0:00:00.825)       1:11:59.783 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nThursday 26 March 2026  06:14:36 +0000 (0:00:00.830)       1:12:00.613 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nThursday 26 March 2026  06:14:37 +0000 (0:00:00.908)       1:12:01.521 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nThursday 26 March 2026  06:14:38 +0000 (0:00:01.113)       1:12:02.635 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nThursday 26 March 2026  06:14:40 +0000 (0:00:01.068)       1:12:03.703 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nThursday 26 March 2026  06:14:41 +0000 (0:00:01.058)       1:12:04.762 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nThursday 26 March 2026  06:14:41 +0000 (0:00:00.822)       1:12:05.584 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nThursday 26 March 2026  06:14:43 +0000 (0:00:01.579)       1:12:07.164 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nThursday 26 March 2026  06:14:44 +0000 (0:00:01.420)       1:12:08.584 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:14:46 +0000 (0:00:01.146)       1:12:09.731 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:14:49 +0000 (0:00:03.253)       1:12:12.984 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nThursday 26 March 2026  06:14:51 +0000 (0:00:01.923)       1:12:14.907 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nThursday 26 March 2026  06:14:52 +0000 (0:00:00.832)       1:12:15.740 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nThursday 26 March 2026  06:14:53 +0000 (0:00:01.128)       1:12:16.868 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nThursday 26 March 2026  06:14:54 +0000 (0:00:01.593)       1:12:18.462 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nThursday 26 March 2026  06:15:00 +0000 (0:00:05.203)       1:12:23.665 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nThursday 26 March 2026  06:15:03 +0000 (0:00:03.118)       1:12:26.784 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nThursday 26 March 2026  06:15:04 +0000 (0:00:01.764)       1:12:28.548 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nThursday 26 March 2026  06:15:06 +0000 (0:00:01.134)       1:12:29.682 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nThursday 26 March 2026  06:15:07 +0000 (0:00:01.604)       1:12:31.287 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nThursday 26 March 2026  06:15:09 +0000 (0:00:01.619)       1:12:32.906 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nThursday 26 March 2026  06:15:39 +0000 (0:00:30.561)       1:13:03.468 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nThursday 26 March 2026  06:15:40 +0000 (0:00:00.768)       1:13:04.237 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nThursday 26 March 2026  06:15:41 +0000 (0:00:00.775)       1:13:05.012 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nThursday 26 March 2026  06:15:42 +0000 (0:00:01.115)       1:13:06.128 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nThursday 26 March 2026  06:15:43 +0000 (0:00:01.109)       1:13:07.237 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nThursday 26 March 2026  06:15:45 +0000 (0:00:02.032)       1:13:09.270 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nThursday 26 March 2026  06:15:47 +0000 (0:00:02.034)       1:13:11.304 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nThursday 26 March 2026  06:15:49 +0000 (0:00:02.238)       1:13:13.543 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nPLAY [Upgrade ceph rbd mirror node] ********************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Upgrade ceph nfs node] ***************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Upgrade ceph client node] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Upgrade ceph-crash daemons] **********************************************\n\nTASK [Stop the ceph-crash service] *********************************************\nThursday 26 March 2026  06:15:54 +0000 (0:00:04.262)       1:13:17.806 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Mask and disable the ceph-crash service] *********************************\nThursday 26 March 2026  06:15:56 +0000 (0:00:02.563)       1:13:20.369 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:16:00 +0000 (0:00:03.343)       1:13:23.713 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:16:02 +0000 (0:00:02.461)       1:13:26.174 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nThursday 26 March 2026  06:16:04 +0000 (0:00:02.407)       1:13:28.582 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nThursday 26 March 2026  06:16:07 +0000 (0:00:02.093)       1:13:30.676 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nThursday 26 March 2026  06:16:09 +0000 (0:00:02.227)       1:13:32.903 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nThursday 26 March 2026  06:16:11 +0000 (0:00:02.322)       1:13:35.226 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nThursday 26 March 2026  06:16:13 +0000 (0:00:02.097)       1:13:37.323 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nThursday 26 March 2026  06:16:16 +0000 (0:00:02.526)       1:13:39.850 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nThursday 26 March 2026  06:16:18 +0000 (0:00:02.109)       1:13:41.960 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nThursday 26 March 2026  06:16:20 +0000 (0:00:02.500)       1:13:44.461 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nThursday 26 March 2026  06:16:22 +0000 (0:00:01.845)       1:13:46.307 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nThursday 26 March 2026  06:16:24 +0000 (0:00:01.755)       1:13:48.062 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nThursday 26 March 2026  06:16:26 +0000 (0:00:02.545)       1:13:50.608 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nThursday 26 March 2026  06:16:28 +0000 (0:00:02.034)       1:13:52.643 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nThursday 26 March 2026  06:16:31 +0000 (0:00:02.197)       1:13:54.841 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nThursday 26 March 2026  06:16:32 +0000 (0:00:01.813)       1:13:56.654 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nThursday 26 March 2026  06:16:35 +0000 (0:00:02.039)       1:13:58.694 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nThursday 26 March 2026  06:16:36 +0000 (0:00:01.825)       1:14:00.520 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nThursday 26 March 2026  06:16:38 +0000 (0:00:02.062)       1:14:02.582 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nThursday 26 March 2026  06:16:40 +0000 (0:00:01.830)       1:14:04.412 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nThursday 26 March 2026  06:16:42 +0000 (0:00:02.115)       1:14:06.528 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nThursday 26 March 2026  06:16:44 +0000 (0:00:01.837)       1:14:08.365 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nThursday 26 March 2026  06:16:46 +0000 (0:00:01.850)       1:14:10.215 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nThursday 26 March 2026  06:16:48 +0000 (0:00:02.318)       1:14:12.533 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nThursday 26 March 2026  06:16:51 +0000 (0:00:03.045)       1:14:15.579 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nThursday 26 March 2026  06:16:55 +0000 (0:00:03.509)       1:14:19.088 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nThursday 26 March 2026  06:16:57 +0000 (0:00:02.520)       1:14:21.608 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nThursday 26 March 2026  06:17:00 +0000 (0:00:02.064)       1:14:23.672 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nThursday 26 March 2026  06:17:02 +0000 (0:00:02.694)       1:14:26.367 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nThursday 26 March 2026  06:17:05 +0000 (0:00:02.909)       1:14:29.276 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Complete upgrade] ********************************************************\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:17:10 +0000 (0:00:04.816)       1:14:34.093 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:17:12 +0000 (0:00:01.712)       1:14:35.806 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Container | disallow pre-reef OSDs and enable all new reef-only functionality] ***\nThursday 26 March 2026  06:17:13 +0000 (0:00:01.604)       1:14:37.410 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Non container | disallow pre-reef OSDs and enable all new reef-only functionality] ***\nThursday 26 March 2026  06:17:16 +0000 (0:00:02.350)       1:14:39.761 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Upgrade node-exporter] ***************************************************\n\nTASK [Stop node-exporter] ******************************************************\nThursday 26 March 2026  06:17:18 +0000 (0:00:01.941)       1:14:41.702 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  06:17:20 +0000 (0:00:02.558)       1:14:44.261 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-engine : Include pre_requisites/prerequisites.yml] ********\nThursday 26 March 2026  06:17:23 +0000 (0:00:02.581)       1:14:46.843 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-container-common : Container registry authentication] ***************\nThursday 26 March 2026  06:17:25 +0000 (0:00:02.498)       1:14:49.341 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-node-exporter : Include setup_container.yml] ************************\nThursday 26 March 2026  06:17:28 +0000 (0:00:02.578)       1:14:51.920 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Upgrade monitoring node] *************************************************\n\nTASK [Stop monitoring services] ************************************************\nThursday 26 March 2026  06:17:31 +0000 (0:00:02.982)       1:14:54.903 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=alertmanager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=prometheus) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=grafana-server) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addr fact - ipv4] ************************\nThursday 26 March 2026  06:17:32 +0000 (0:00:01.282)       1:14:56.185 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addr fact - ipv6] ************************\nThursday 26 March 2026  06:17:33 +0000 (0:00:01.158)       1:14:57.344 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addrs fact - ipv4] ***********************\nThursday 26 March 2026  06:17:34 +0000 (0:00:01.175)       1:14:58.520 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addrs fact - ipv6] ***********************\nThursday 26 March 2026  06:17:35 +0000 (0:00:01.124)       1:14:59.645 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Create prometheus directories] *************************\nThursday 26 March 2026  06:17:37 +0000 (0:00:01.176)       1:15:00.821 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/prometheus) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/var/lib/prometheus) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Write prometheus config file] **************************\nThursday 26 March 2026  06:17:38 +0000 (0:00:01.121)       1:15:01.943 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Make sure the alerting rules directory exists] *********\nThursday 26 March 2026  06:17:39 +0000 (0:00:01.148)       1:15:03.091 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Copy alerting rules] ***********************************\nThursday 26 March 2026  06:17:40 +0000 (0:00:01.112)       1:15:04.204 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Create alertmanager directories] ***********************\nThursday 26 March 2026  06:17:41 +0000 (0:00:01.343)       1:15:05.548 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/alertmanager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/var/lib/alertmanager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Write alertmanager config file] ************************\nThursday 26 March 2026  06:17:43 +0000 (0:00:01.136)       1:15:06.684 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-prometheus : Include setup_container.yml] ***************************\nThursday 26 March 2026  06:17:44 +0000 (0:00:01.103)       1:15:07.788 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-grafana : Include setup_container.yml] ******************************\nThursday 26 March 2026  06:17:45 +0000 (0:00:01.103)       1:15:08.891 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [ceph-grafana : Include configure_grafana.yml] ****************************\nThursday 26 March 2026  06:17:46 +0000 (0:00:01.116)       1:15:10.008 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Upgrade ceph dashboard] **************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nThursday 26 March 2026  06:17:47 +0000 (0:00:01.641)       1:15:11.649 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addr fact - ipv4] ************************\nThursday 26 March 2026  06:17:49 +0000 (0:00:01.793)       1:15:13.443 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addr fact - ipv6] ************************\nThursday 26 March 2026  06:17:51 +0000 (0:00:01.527)       1:15:14.970 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addrs fact - ipv4] ***********************\nThursday 26 March 2026  06:17:52 +0000 (0:00:01.430)       1:15:16.401 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set grafana_server_addrs fact - ipv6] ***********************\nThursday 26 March 2026  06:17:54 +0000 (0:00:01.759)       1:15:18.160 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-dashboard : Include configure_dashboard.yml] ************************\nThursday 26 March 2026  06:17:55 +0000 (0:00:01.437)       1:15:19.597 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-dashboard : Print dashboard URL] ************************************\nThursday 26 March 2026  06:17:57 +0000 (0:00:01.397)       1:15:20.995 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Switch any existing crush buckets to straw2] *****************************\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nThursday 26 March 2026  06:17:59 +0000 (0:00:02.089)       1:15:23.085 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nThursday 26 March 2026  06:18:00 +0000 (0:00:01.526)       1:15:24.611 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set_fact ceph_cmd] *******************************************************\nThursday 26 March 2026  06:18:02 +0000 (0:00:01.243)       1:15:25.854 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Backup the crushmap] *****************************************************\nThursday 26 March 2026  06:18:03 +0000 (0:00:01.114)       1:15:26.969 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Switch crush buckets to straw2] ******************************************\nThursday 26 March 2026  06:18:06 +0000 (0:00:02.960)       1:15:29.929 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Remove crushmap backup] **************************************************\nThursday 26 March 2026  06:18:09 +0000 (0:00:03.295)       1:15:33.225 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Show ceph status] ********************************************************\n\nTASK [Set_fact container_exec_cmd_status] **************************************\nThursday 26 March 2026  06:18:11 +0000 (0:00:01.963)       1:15:35.189 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Show ceph status] ********************************************************\nThursday 26 March 2026  06:18:13 +0000 (0:00:02.269)       1:15:37.458 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Show all daemons version] ************************************************\nThursday 26 March 2026  06:18:16 +0000 (0:00:02.338)       1:15:39.797 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : ok=0    changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=76  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=248 \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=376 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=191 \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=350 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=196 \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=351 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=317 \u001b[0m \u001b[0;33mchanged=21  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=362 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=307 \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=359 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=303 \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=344 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  06:18:19 +0000 (0:00:03.761)       1:15:43.559 ******** \n=============================================================================== \nDisable pg autoscale on pools ------------------------------------------ 74.54s\nRe-enable pg autoscale on pools ---------------------------------------- 74.51s\nGather and delegate facts ---------------------------------------------- 31.50s\nceph-rgw : Create rgw pools -------------------------------------------- 31.45s\nceph-rgw : Create rgw pools -------------------------------------------- 31.26s\nceph-rgw : Create rgw pools -------------------------------------------- 30.56s\nWaiting for clean pgs... ----------------------------------------------- 29.82s\nceph-mon : Set cluster configs ----------------------------------------- 28.66s\nceph-config : Set config to cluster ------------------------------------ 23.07s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 23.01s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 22.82s\nStop ceph mgr ---------------------------------------------------------- 17.87s\nCreate potentially missing keys (rbd and rbd-mirror) ------------------- 16.74s\nceph-osd : Wait for all osd to be up ----------------------------------- 15.47s\nceph-config : Set config to cluster ------------------------------------ 14.42s\nceph-config : Set osd_memory_target to cluster host config ------------- 12.82s\nceph-config : Set osd_memory_target to cluster host config ------------- 12.81s\nStop ceph osd ---------------------------------------------------------- 12.80s\nStop standby ceph mds -------------------------------------------------- 11.17s\nStop ceph mon ---------------------------------------------------------- 10.37s\n+ osism apply cephclient\n\u001b[32m2026-03-26 06:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21ea1124-23b4-48c3-93da-984335f8d483 (cephclient) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-26 06:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 21ea1124-23b4-48c3-93da-984335f8d483 (cephclient) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x79be164d1be0>): Expecting value: line 2 column 1 (char 1)\u001b[0m\n\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m\n\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m\n\u001b[1;35m0x79be164d1be0>): 'NoneType' object is not subscriptable\u001b[0m\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nThursday 26 March 2026  06:18:31 +0000 (0:00:02.472)       0:00:02.472 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nThursday 26 March 2026  06:18:32 +0000 (0:00:00.801)       0:00:03.273 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nThursday 26 March 2026  06:18:33 +0000 (0:00:01.721)       0:00:04.995 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nThursday 26 March 2026  06:18:34 +0000 (0:00:01.120)       0:00:06.116 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nThursday 26 March 2026  06:18:35 +0000 (0:00:00.932)       0:00:07.049 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nThursday 26 March 2026  06:18:36 +0000 (0:00:00.951)       0:00:08.001 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nThursday 26 March 2026  06:18:38 +0000 (0:00:01.275)       0:00:09.276 ******** \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nThursday 26 March 2026  06:18:42 +0000 (0:00:04.160)       0:00:13.437 ******** \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nThursday 26 March 2026  06:18:42 +0000 (0:00:00.508)       0:00:13.946 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nThursday 26 March 2026  06:18:42 +0000 (0:00:00.152)       0:00:14.098 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=8   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nThursday 26 March 2026  06:18:44 +0000 (0:00:01.113)       0:00:15.211 ******** \n=============================================================================== \nosism.services.cephclient : Copy wrapper scripts ------------------------ 4.16s\nosism.services.cephclient : Create required directories ----------------- 1.72s\nosism.services.cephclient : Manage cephclient service ------------------- 1.28s\nosism.services.cephclient : Copy configuration files -------------------- 1.12s\nosism.services.cephclient : Include rook task --------------------------- 1.11s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 0.95s\nosism.services.cephclient : Copy keyring file --------------------------- 0.93s\nosism.services.cephclient : Include container tasks --------------------- 0.80s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s\nosism.services.cephclient : Include package tasks ----------------------- 0.15s\n+ [[ false == \\f\\a\\l\\s\\e ]]\n+ sh -c /opt/configuration/scripts/upgrade/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=9.5.0\n++ MANAGER_VERSION=9.5.0\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=false\n++ TEMPEST=false\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n++ export RABBITMQ3TO4=true\n++ RABBITMQ3TO4=true\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n++ export MANAGER_VERSION=10.0.0-rc.1\n++ MANAGER_VERSION=10.0.0-rc.1\n+ [[ true == \\t\\r\\u\\e ]]\n+ osism migrate rabbitmq3to4 prepare\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to RabbitMQ Management API at 192.168.16.10:15672 (node: testbed-node-0) as openstack...\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mKolla configuration check passed\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreated vhost 'openstack' with default_queue_type=quorum\u001b[0m\n\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSet permissions for user 'openstack' on vhost 'openstack'\u001b[0m\n+ osism migrate rabbitmq3to4 list\n\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to RabbitMQ Management API at 192.168.16.10:15672 (node: testbed-node-0) as openstack...\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFound 208 classic queue(s) in vhost '/':\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - alarm.all.sample (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - alarming.sample (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers.barbican.queue (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_b4e210291c1f4595a8110eea41f3b718 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_e5817fbc18e24736ae19aea2dd46443c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_fb5c5d01a8bc44969546e21abf2484cc (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican_notifications.info (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_35733cd31394488d9e3db0d42e40e34c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_3c6a3e144f2243b2bc8f0af87b8fdf99 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_3e97b3e6bc4c46eab56fa9a43109ebd7 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_6bd3253d59d0488c8ecbab32dee32db2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_82cf3e0940194386804159a753c7460c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_d4f9292219c84761a99423dd67f1398c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_8f5f559e775243a98c2c1f750404a4f1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_bf6b6bf9a62148dea1244fab3e97b501 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_f2d658f5fc36455a8931e7580182289c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_7eb0367bc74f44d2b4625e36bc5d642d (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_d047f442bc244458a150b5609e4af3ec (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_f872bbe8a39c496c82652ebaf402b14a (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes_fanout_a6533171c6994adf9bcce0cd216bf57c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes_fanout_b4bcfca53a99458ab09955313468afc6 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes_fanout_ade417b8c94e4319896a910815b50d48 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_029b93b2eb2647bebe3b03dbcaf16cff (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_8c9dd967d35c4794ac8b90812623d27b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_8fd2c7d0d57d479098d7dc9f2a8a7324 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-3 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-4 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-5 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_1633d9b6ca8d45e3980c3aa8b6d1479f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_2ed9f7af58614a9fac21fe0ebf645431 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_dce43fa4b04f4a428d9422d5117ae051 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_0a087e21fbc9400eaca4d0d8c28e18e1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_2779a7fe25ce4c3292b21b1ec783f5a8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_3807aa7d35c84957b24855e76e40df5c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_39ddbc7364ef4e6093e49dc1f29da98e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_3a1d92acabb3468e9eeb34fd1571f679 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_434589943aa54e9c8edd5825fd358e9e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - event.sample (vhost: /, messages: 9)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.dxt76xirij65 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.u74qoctmj2at (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.ylgl7fnvnb3q (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_114f868d641941c393c267ebefb2d881 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_25275d6376434f9a88947423f3f0ddf9 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_4ada99082e6346f38692f45c23a45fb8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_4b77c13e89a54dfdab9272dd2f0ef97f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_a7c5e48d18c443a1956b99c2edbcbc94 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_b3b098cb100849bab05884917ef2e4e2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_e4ecb03c594f47fe9337c88ff65131bd (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_ee984a30032243dfb0cc80ab0c2b8384 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_f4be591b6d034162b990cb9296faba3f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_c0ecdb1bd3924f28b9e1bbc7eccc646b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_f03d3befe0314d879d7f796de934dd9d (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_f85984553bde4d31bf278d156c7ac211 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_a136d7d13b474010a342264c2c907e16 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_e2eb07a671e64e1686089b8a01af098e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_f7be0f8d31c34c4da429bfe2aca47847 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-0@cephfsnative1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-1@cephfsnative1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-2@cephfsnative1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_7805b1c8e1414693b0ba1f609b02809a (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_bc63684f7fde4834a4efc8c16b403a68 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_fc5e9c2e25184fcfb3dfda7a4dc4dc33 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.audit (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.critical (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.debug (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.error (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.info (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.sample (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.warn (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_0fac4d064de84e5bbd6d466dd834e1f4 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_86ab57fd165d4bd783e3c86dd75f8847 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_c60f129848694c27b399ca7ddbf06f06 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_1598d067d87e4d04b00c428b4b47727f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_36061f07216b4217adb49fe71c35eaf8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_43e908b0d0d945869d271173a5152aba (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_ae6a2357e6a34a50b6620a7e1a181a77 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_cbd89788cf2f40d2a554ea5281955327 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_e0f7d2d7fb274163a86be0a37dfd2cd8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_44a4c0e6ec6844979a126b54345b60f3 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_63aafcca795041abb90dcd32d46b2b55 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_80382bcbeca9481f812885405d457324 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_a296e3d4085d407584dbef3c971a0f2a (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_b4cf746c93324effb7ff69cab9c2350a (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_c3871e367a2d4d3baf01c5aea2c1919a (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f41bb944c69d473ab16fa6b5d6d35e10 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f7d147aad30f4bcb9adb25eeb33c86c0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f9b48b37b74d417d94f841025641bd73 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_1f4456cdaa8043e3be2805ad39d5ab25 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_26775edc33404dbc9abe1e144bd80525 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_31a15587650e47c3bde5abc053e22c01 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_32ec2ee87c0245788dffd816ea3c023c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_4a6c18d1e8814a9dae92e88643a4f6a3 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_538ba911017345b08e67db1a5c5b55b5 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_5897b7a973ad4acc91343aa05c7af3ff (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6319d5e241c541e58310e16e9dc52144 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6b360b1659a14d0da13ea3b346e563aa (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6f16a59b15654f4dbcf85abee6644a2b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_94ee497291ae489a9c6b238ca80e38fd (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_9e52c6dc9ed24a7b80ea13a94674ad74 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_aa918d06050f4bd7a2ee224ba419ce31 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_c1cdb63a10e04fc6bf41640f72fa8159 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_cfb3d53116bd463396bf620650870a65 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_d09b359918574dcab62f2ee6b5dcdca8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_dad219db8f49450d9a2b986223f703bf (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_dd786927624c419ead6015dcfc58770f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_1cb9224fc37a4a48bfacc80953dd403f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_3135493c2b264f71bb9bbad34fb00452 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_3d6dfb328ccc46a6a9cb2f4b96e3d697 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_408591ee4ea44bc4bacb6ad94f1cf98e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_4500faec64a44eedbcb004f0e67b9b20 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_64da208a82c3494385ded2e418c9a122 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_651674cdba1c43a9813d7ecb595a9cfb (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_c6e201d63611428cad0b60e848404955 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_fa758238da8947f4b71f8bcdaabb883b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_0143f57daf2b426b89201cef986f95e2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_1607a4c70ae045398f71fc2d451c5a27 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_2d05cc2aedda45ea9c31797c9488552e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_2df21c6aa5e242c2a4bd19ebf039c414 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_42c674747649424c96f46572b03ea0da (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_4ef0c8bf75cf449ca8256621ad7454ac (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_550a4fdad56e49009b36c1374a865c76 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_919bd19646db497b8240f22dec5e8b0c (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_91fe99c6abc64314ada5daa43f18f50b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_9db8e0e8fdb44324a47d8bfc1ac1f280 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a0bf128c59ef4dcf98d30d7684f34f2e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a1d93d6f81c749c9b16200e2c0c20728 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a363e6378e634bcaaa24df418226ca2b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_d2aaa10c347d4f878af4a93b4462aca8 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_dcf1dff11157461c91e0f07e1149d443 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_e5a2bc26f56f421d9a8b1028dfa744bf (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_e6549370496a4080a6647b33c1437c75 (vhost: /, messages: 1)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_ece0d2b5146749299699600bd6620114 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_ee22625ab8d44350bdd04b376e72d57f (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_17c056e993ef47e3acc112af16d457e3 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_495244cec2c64f3c9bdf3534c4821ed3 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_4f6c87c162d1459ea749bf93d9575641 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_5825e34b1d114c22a0ea70626d5b3c06 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_9dd488cc166144d3b66b6167500e5b69 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_d8d2fe93979a46f4ad96c7e79aad00d7 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-0 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-1 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-2 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_360ce71e2a2b43c898e071ad25434240 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_4b3c4dcb71b947008464d6f8006593a6 (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_4f0de0701f684613b3655e8683bea71b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_5a8bae1580fd4d94b676dd0da9c1331b (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_c08facaf21194343b22a11cadc24a44e (vhost: /, messages: 0)\u001b[0m\n\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_cc85c421d64b4d6b9fbd61b01ede5463 (vhost: /, messages: 0)\u001b[0m\n+ osism migrate rabbitmq3to4 list-exchanges\nusage: osism migrate rabbitmq3to4 [-h] [--server SERVER] [--dry-run]\n                                  [--no-close-connections] [--quorum]\n                                  [--vhost VHOST]\n                                  [{list,delete,prepare,check}]\n                                  [{aodh,barbican,ceilometer,cinder,designate,notifications,manager,magnum,manila,neutron,nova,octavia}]\nosism migrate rabbitmq3to4: error: argument command: invalid choice: 'list-exchanges' (choose from list, delete, prepare, check)",
                            "stdout_lines": [
                                "+ set -e",
                                "+ export MANAGER_VERSION=10.0.0-rc.1",
                                "+ MANAGER_VERSION=10.0.0-rc.1",
                                "+ CEPH_VERSION=reef",
                                "+ OPENSTACK_VERSION=2024.2",
                                "+ KOLLA_NAMESPACE=kolla/release",
                                "+ sh -c '/opt/configuration/scripts/upgrade-manager.sh 10.0.0-rc.1 reef 2024.2 kolla/release'",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "++ docker inspect --format '{{ index .Config.Labels \"org.opencontainers.image.version\"}}' osism-ansible",
                                "+ OLD_MANAGER_VERSION=v0.20251130.0",
                                "++ docker inspect --format '{{ index .Config.Labels \"de.osism.release.openstack\"}}' kolla-ansible",
                                "",
                                "# UPGRADE MANAGER",
                                "",
                                "+ OLD_OPENSTACK_VERSION=2024.2",
                                "+ echo",
                                "+ echo '# UPGRADE MANAGER'",
                                "+ echo",
                                "+ export MANAGER_VERSION=10.0.0-rc.1",
                                "+ MANAGER_VERSION=10.0.0-rc.1",
                                "+ CEPH_VERSION=reef",
                                "+ OPENSTACK_VERSION=2024.2",
                                "+ KOLLA_NAMESPACE=kolla/release",
                                "+ /opt/configuration/scripts/set-manager-version.sh 10.0.0-rc.1",
                                "+ set -e",
                                "+ VERSION=10.0.0-rc.1",
                                "+ sed -i 's/manager_version: .*/manager_version: 10.0.0-rc.1/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ 10.0.0-rc.1 != \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i /ceph_version:/d /opt/configuration/environments/manager/configuration.yml",
                                "+ sed -i /openstack_version:/d /opt/configuration/environments/manager/configuration.yml",
                                "+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh",
                                "/opt/configuration ~",
                                "+ set -e",
                                "+ pushd /opt/configuration",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging",
                                "Requirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)",
                                "Requirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)",
                                "Requirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)",
                                "Requirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)",
                                "Requirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)",
                                "Requirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)",
                                "Requirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)",
                                "Requirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)",
                                "Requirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)",
                                "Requirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)",
                                "Requirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)",
                                "Requirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)",
                                "Requirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)",
                                "Requirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)",
                                "++ which gilt",
                                "+ GILT=/opt/venv/bin/gilt",
                                "+ /opt/venv/bin/gilt overlay",
                                "osism.cfg-generics:",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/",
                                "  - running `rm render-images.py` in /opt/configuration/environments/manager/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/",
                                "  - running `rm set-versions.py` in /opt/configuration/environments/",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ popd",
                                "~",
                                "+ [[ 10.0.0-rc.1 == \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release",
                                "+ set -e",
                                "+ NAMESPACE=kolla/release",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "+ sh -c /opt/configuration/scripts/sync-configuration-repository.sh",
                                "/opt/configuration ~",
                                "+ set -e",
                                "+ pushd /opt/configuration",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging",
                                "Requirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3)",
                                "Requirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.33.0)",
                                "Requirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6)",
                                "Requirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.3)",
                                "Requirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (26.0)",
                                "Requirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.3.1)",
                                "Requirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6)",
                                "Requirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20)",
                                "Requirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2)",
                                "Requirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.6)",
                                "Requirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.11)",
                                "Requirement already satisfied: urllib3<3,>=1.26 in /opt/venv/lib/python3.12/site-packages (from requests) (2.6.3)",
                                "Requirement already satisfied: certifi>=2023.5.7 in /opt/venv/lib/python3.12/site-packages (from requests) (2026.2.25)",
                                "Requirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.3)",
                                "++ which gilt",
                                "+ GILT=/opt/venv/bin/gilt",
                                "+ /opt/venv/bin/gilt overlay",
                                "osism.cfg-generics:",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/",
                                "  - copied (v0.20251130.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/",
                                "  - running `rm render-images.py` in /opt/configuration/environments/manager/",
                                "  - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/",
                                "  - running `rm set-versions.py` in /opt/configuration/environments/",
                                "~",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ popd",
                                "++ semver v0.20251130.0 6.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 10.0.0-rc.1 10.0.0-0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ sed -i '/^om_enable_rabbitmq_high_availability:/d' /opt/configuration/environments/kolla/configuration.yml",
                                "+ sed -i '/^om_enable_rabbitmq_quorum_queues:/d' /opt/configuration/environments/kolla/configuration.yml",
                                "+++ semver v0.20251130.0 9.5.0",
                                "++ '[' -1 -le 0 ']'",
                                "+++ semver 10.0.0-rc.1 10.0.0-0",
                                "++ '[' 1 -ge 0 ']'",
                                "++ echo true",
                                "+ MANAGER_UPGRADE_CROSSES_10=true",
                                "+++ semver 2024.2 2024.2",
                                "++ '[' 0 -le 0 ']'",
                                "+++ semver 2024.2 2025.1",
                                "++ '[' -1 -ge 0 ']'",
                                "++ echo false",
                                "+ OPENSTACK_UPGRADE_CROSSES_2025=false",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ echo 'om_rpc_vhost: openstack'",
                                "+ echo 'om_notify_vhost: openstack'",
                                "+ sed -i 's#manager_listener_broker_vhost: .*#manager_listener_broker_vhost: /openstack#g' /opt/configuration/environments/manager/configuration.yml",
                                "+ echo 'export RABBITMQ3TO4=true'",
                                "+ sudo tee -a /opt/manager-vars.sh",
                                "export RABBITMQ3TO4=true",
                                "+ osism update manager",
                                "Collecting uv",
                                "  Downloading uv-0.11.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB)",
                                "Downloading uv-0.11.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (24.5 MB)",
                                "   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 24.5/24.5 MB 33.8 MB/s eta 0:00:00",
                                "Installing collected packages: uv",
                                "Successfully installed uv-0.11.1",
                                "Resolved 11 packages in 380ms",
                                "Downloading netaddr (2.2MiB)",
                                "Downloading cryptography (4.3MiB)",
                                "Downloading ansible-core (2.1MiB)",
                                "Downloading ansible (54.5MiB)",
                                " Downloaded netaddr",
                                " Downloaded cryptography",
                                " Downloaded ansible-core",
                                " Downloaded ansible",
                                "Prepared 11 packages in 7.40s",
                                "Installed 11 packages in 604ms",
                                " + ansible==11.11.0",
                                " + ansible-core==2.18.15",
                                " + cffi==2.0.0",
                                " + cryptography==46.0.6",
                                " + jinja2==3.1.6",
                                " + markupsafe==3.0.3",
                                " + netaddr==1.3.0",
                                " + packaging==26.0",
                                " + pycparser==3.0",
                                " + pyyaml==6.0.3",
                                " + resolvelib==1.0.1",
                                "Cloning into '/home/dragon/.ansible/tmp/ansible-local-203816n2qsv95p/tmpz7vqf5eh/ansible-collection-servicesarne1aad'...",
                                "Your branch is up to date with 'origin/main'.",
                                "Already on 'main'",
                                "Starting galaxy collection install process",
                                "Process install dependency map",
                                "Starting collection install process",
                                "Installing 'osism.services:999.0.0' to '/home/dragon/.ansible/collections/ansible_collections/osism/services'",
                                "Created collection for osism.services:999.0.0 at /home/dragon/.ansible/collections/ansible_collections/osism/services",
                                "osism.services:999.0.0 was installed successfully",
                                "Cloning into '/home/dragon/.ansible/tmp/ansible-local-203833xiam2p3s/tmp84azf_u6/ansible-playbooks-managerqcge2toh'...",
                                "Already on 'main'",
                                "Your branch is up to date with 'origin/main'.",
                                "Starting galaxy collection install process",
                                "Process install dependency map",
                                "Starting collection install process",
                                "Installing 'osism.manager:999.0.0' to '/home/dragon/.ansible/collections/ansible_collections/osism/manager'",
                                "Created collection for osism.manager:999.0.0 at /home/dragon/.ansible/collections/ansible_collections/osism/manager",
                                "osism.manager:999.0.0 was installed successfully",
                                "[WARNING]: Invalid characters were found in group names but not replaced, use",
                                "-vvvv to see details",
                                "",
                                "PLAY [Apply role manager] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "ok: [testbed-manager] => (item=/opt/ansible)",
                                "ok: [testbed-manager] => (item=/opt/archive)",
                                "ok: [testbed-manager] => (item=/opt/manager/configuration)",
                                "ok: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "ok: [testbed-manager] => (item=/opt/manager/secrets)",
                                "ok: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "ok: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "ok: [testbed-manager] => (item=ara)",
                                "ok: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "ok: [testbed-manager] => (item=None)",
                                "ok: [testbed-manager] => (item=None)",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "ok: [testbed-manager] => (item=conductor)",
                                "ok: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "ok: [testbed-manager] => (item=osism)",
                                "ok: [testbed-manager] => (item=osism-update-docker)",
                                "ok: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "ok: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Register that manager service was restarted] ***",
                                "ok: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /home/dragon/.ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:0.20251208.0\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=51   changed=9    unreachable=0    failed=0    skipped=8    rescued=0    ignored=0   ",
                                "",
                                "\u001b[32m2026-03-26 04:20:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c34b26f2-f289-4f95-835a-bf861cbcec99 (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "+ [[ '' == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                         COMMAND                  SERVICE                CREATED         STATUS                          PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:0.20251208.0           \"/entrypoint.sh osis\u2026\"   ceph-ansible           3 minutes ago   Up 2 minutes (healthy)          ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:0.20251208.0          \"/entrypoint.sh osis\u2026\"   kolla-ansible          3 minutes ago   Up 2 minutes (healthy)          ",
                                "manager-api-1                    registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   api                    3 minutes ago   Up 3 minutes (healthy)          192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                    \"sh -c '/wait && /ru\u2026\"   ara-server             2 hours ago     Up 2 minutes (healthy)          8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   beat                   3 minutes ago   Up 3 minutes (healthy)          ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   flower                 3 minutes ago   Up 3 minutes (healthy)          ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:0.20251208.0   \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   3 minutes ago   Up 2 minutes (healthy)          ",
                                "manager-listener-1               registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   listener               3 minutes ago   Restarting (0) 21 seconds ago   ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4          \"docker-entrypoint.s\u2026\"   mariadb                2 hours ago     Up 3 minutes (healthy)          3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- osism\u2026\"   openstack              3 minutes ago   Up 3 minutes (healthy)          ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine      \"docker-entrypoint.s\u2026\"   redis                  2 hours ago     Up 3 minutes (healthy)          6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:0.20251208.0          \"/entrypoint.sh osis\u2026\"   osism-ansible          3 minutes ago   Up 2 minutes (healthy)          ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:0.20251208.0         \"docker-entrypoint.s\u2026\"   frontend               3 minutes ago   Up 3 minutes                    192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:0.20251208.0       \"/entrypoint.sh osis\u2026\"   osism-kubernetes       3 minutes ago   Up 2 minutes (healthy)          ",
                                "osismclient                      registry.osism.tech/osism/osism:0.20251208.0                  \"/sbin/tini -- sleep\u2026\"   osismclient            3 minutes ago   Up 3 minutes (healthy)          ",
                                "+ [[ '' == \\t\\r\\u\\e ]]",
                                "+ [[ '' == \\f\\a\\l\\s\\e ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-03-26 04:20:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe19af64-9cef-441d-8bc1-20715ae45ad3 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:20:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fe19af64-9cef-441d-8bc1-20715ae45ad3 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Thursday 26 March 2026  04:20:49 +0000 (0:00:01.971)       0:00:01.971 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Thursday 26 March 2026  04:20:52 +0000 (0:00:03.524)       0:00:05.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Thursday 26 March 2026  04:20:55 +0000 (0:00:02.580)       0:00:08.077 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Thursday 26 March 2026  04:21:02 +0000 (0:00:07.088)       0:00:15.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:21:05 +0000 (0:00:02.682)       0:00:17.848 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 7.09s",
                                "osism.commons.facts : Create custom facts directory --------------------- 3.53s",
                                "Gather facts for all hosts ---------------------------------------------- 2.68s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 2.58s",
                                "++ semver 10.0.0-rc.1 10.0.0-0",
                                "+ [[ 1 -ge 0 ]]",
                                "++ docker inspect --format '{{ index .Config.Labels \"de.osism.release.openstack\"}}' kolla-ansible",
                                "+ OPENSTACK_VERSION=2025.1",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla/release/2025.1",
                                "+ set -e",
                                "+ NAMESPACE=kolla/release/2025.1",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla/release/2025.1#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "+ sh -c /opt/configuration/scripts/upgrade-services.sh",
                                "",
                                "# UPGRADE SERVICES",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# UPGRADE SERVICES'",
                                "+ echo",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "++ export RABBITMQ3TO4=true",
                                "++ RABBITMQ3TO4=true",
                                "+ SKIP_OPENSTACK_UPGRADE=false",
                                "+ SKIP_CEPH_UPGRADE=false",
                                "+ sh -c /opt/configuration/scripts/pull-images.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "",
                                "# PULL IMAGES",
                                "",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "++ export RABBITMQ3TO4=true",
                                "++ RABBITMQ3TO4=true",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "+ echo",
                                "++ semver 9.5.0 7.0.0",
                                "+ [[ 1 -ge 0 ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-03-26 04:21:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-03-26 04:21:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 411ad312-6a16-41aa-83a0-93c728c3fae0 (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:21:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 411ad312-6a16-41aa-83a0-93c728c3fae0 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/upgrade/500-kubernetes.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "++ export MANAGER_VERSION=10.0.0-rc.1",
                                "++ MANAGER_VERSION=10.0.0-rc.1",
                                "++ semver 10.0.0-rc.1 8.0.3",
                                "+ [[ 1 -ge 0 ]]",
                                "+ osism apply frr",
                                "\u001b[32m2026-03-26 04:21:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d394b63-d017-4309-be34-0efdf0caac8b (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:21:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0d394b63-d017-4309-be34-0efdf0caac8b (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Thursday 26 March 2026  04:21:38 +0000 (0:00:02.965)       0:00:02.965 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Thursday 26 March 2026  04:21:40 +0000 (0:00:02.108)       0:00:05.074 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Thursday 26 March 2026  04:21:42 +0000 (0:00:02.058)       0:00:07.133 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Thursday 26 March 2026  04:21:45 +0000 (0:00:02.810)       0:00:09.943 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Thursday 26 March 2026  04:21:47 +0000 (0:00:01.954)       0:00:11.898 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Thursday 26 March 2026  04:21:49 +0000 (0:00:01.980)       0:00:13.878 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Thursday 26 March 2026  04:21:51 +0000 (0:00:02.413)       0:00:16.291 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Thursday 26 March 2026  04:21:53 +0000 (0:00:01.167)       0:00:17.459 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Thursday 26 March 2026  04:21:54 +0000 (0:00:01.140)       0:00:18.600 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Thursday 26 March 2026  04:21:56 +0000 (0:00:01.916)       0:00:20.516 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Thursday 26 March 2026  04:21:59 +0000 (0:00:03.629)       0:00:24.146 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:22:02 +0000 (0:00:02.615)       0:00:26.761 ******** ",
                                "=============================================================================== ",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 3.63s",
                                "osism.services.frr : Install frr package -------------------------------- 2.81s",
                                "osism.services.frr : Manage frr service --------------------------------- 2.62s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 2.41s",
                                "osism.services.frr : Include distribution specific install tasks -------- 2.11s",
                                "osism.services.frr : Pin frr package version ---------------------------- 2.06s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 1.98s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.95s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.92s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 1.17s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 1.14s",
                                "+ osism apply kubernetes",
                                "\u001b[32m2026-03-26 04:22:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe2e5425-917f-42a5-93e8-ab4ed1c92939 (kubernetes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:22:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fe2e5425-917f-42a5-93e8-ab4ed1c92939 (kubernetes) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Thursday 26 March 2026  04:22:11 +0000 (0:00:01.796)       0:00:01.796 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Thursday 26 March 2026  04:22:16 +0000 (0:00:05.407)       0:00:07.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Thursday 26 March 2026  04:22:18 +0000 (0:00:01.878)       0:00:09.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Thursday 26 March 2026  04:22:20 +0000 (0:00:01.950)       0:00:11.033 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Thursday 26 March 2026  04:22:23 +0000 (0:00:03.154)       0:00:14.187 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Thursday 26 March 2026  04:22:26 +0000 (0:00:02.308)       0:00:16.496 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Thursday 26 March 2026  04:22:28 +0000 (0:00:02.093)       0:00:18.590 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Thursday 26 March 2026  04:22:30 +0000 (0:00:02.232)       0:00:20.822 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Thursday 26 March 2026  04:22:32 +0000 (0:00:01.722)       0:00:22.544 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Thursday 26 March 2026  04:22:34 +0000 (0:00:01.977)       0:00:24.522 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Thursday 26 March 2026  04:22:36 +0000 (0:00:02.055)       0:00:26.577 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Thursday 26 March 2026  04:22:39 +0000 (0:00:03.010)       0:00:29.588 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Thursday 26 March 2026  04:22:41 +0000 (0:00:02.691)       0:00:32.280 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Thursday 26 March 2026  04:22:43 +0000 (0:00:01.934)       0:00:34.214 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Thursday 26 March 2026  04:22:46 +0000 (0:00:02.286)       0:00:36.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Thursday 26 March 2026  04:22:47 +0000 (0:00:01.698)       0:00:38.199 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Thursday 26 March 2026  04:22:49 +0000 (0:00:02.052)       0:00:40.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Thursday 26 March 2026  04:22:51 +0000 (0:00:01.867)       0:00:42.119 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Thursday 26 March 2026  04:22:54 +0000 (0:00:02.859)       0:00:44.979 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Thursday 26 March 2026  04:22:56 +0000 (0:00:01.767)       0:00:46.747 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Thursday 26 March 2026  04:22:58 +0000 (0:00:02.080)       0:00:48.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Thursday 26 March 2026  04:23:00 +0000 (0:00:02.134)       0:00:50.962 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Thursday 26 March 2026  04:23:02 +0000 (0:00:01.964)       0:00:52.926 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Thursday 26 March 2026  04:23:03 +0000 (0:00:01.363)       0:00:54.290 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Thursday 26 March 2026  04:23:05 +0000 (0:00:01.705)       0:00:55.995 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Thursday 26 March 2026  04:23:07 +0000 (0:00:02.200)       0:00:58.196 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Thursday 26 March 2026  04:23:09 +0000 (0:00:01.908)       0:01:00.104 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Thursday 26 March 2026  04:23:12 +0000 (0:00:02.526)       0:01:02.631 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Thursday 26 March 2026  04:23:13 +0000 (0:00:01.669)       0:01:04.300 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Thursday 26 March 2026  04:23:15 +0000 (0:00:01.858)       0:01:06.159 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Thursday 26 March 2026  04:23:18 +0000 (0:00:02.518)       0:01:08.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Thursday 26 March 2026  04:23:19 +0000 (0:00:01.368)       0:01:10.046 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Thursday 26 March 2026  04:23:21 +0000 (0:00:01.699)       0:01:11.745 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Thursday 26 March 2026  04:23:23 +0000 (0:00:02.229)       0:01:13.975 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Thursday 26 March 2026  04:23:25 +0000 (0:00:01.820)       0:01:15.796 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Thursday 26 March 2026  04:23:26 +0000 (0:00:01.411)       0:01:17.208 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Thursday 26 March 2026  04:23:50 +0000 (0:00:23.365)       0:01:40.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Thursday 26 March 2026  04:23:51 +0000 (0:00:01.362)       0:01:41.936 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Thursday 26 March 2026  04:23:53 +0000 (0:00:02.075)       0:01:44.011 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Thursday 26 March 2026  04:23:55 +0000 (0:00:02.314)       0:01:46.326 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Thursday 26 March 2026  04:24:23 +0000 (0:00:27.061)       0:02:13.387 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Thursday 26 March 2026  04:24:24 +0000 (0:00:01.800)       0:02:15.187 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Thursday 26 March 2026  04:24:26 +0000 (0:00:01.708)       0:02:16.895 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Thursday 26 March 2026  04:24:28 +0000 (0:00:01.997)       0:02:18.893 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Thursday 26 March 2026  04:24:30 +0000 (0:00:01.744)       0:02:20.638 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Thursday 26 March 2026  04:24:31 +0000 (0:00:01.384)       0:02:22.023 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Thursday 26 March 2026  04:24:33 +0000 (0:00:01.654)       0:02:23.677 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Thursday 26 March 2026  04:24:35 +0000 (0:00:01.978)       0:02:25.655 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Thursday 26 March 2026  04:24:37 +0000 (0:00:01.865)       0:02:27.520 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Thursday 26 March 2026  04:24:39 +0000 (0:00:02.004)       0:02:29.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Thursday 26 March 2026  04:24:40 +0000 (0:00:01.370)       0:02:30.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Thursday 26 March 2026  04:24:41 +0000 (0:00:01.317)       0:02:32.214 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Thursday 26 March 2026  04:24:43 +0000 (0:00:01.658)       0:02:33.872 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Thursday 26 March 2026  04:24:45 +0000 (0:00:01.650)       0:02:35.523 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Thursday 26 March 2026  04:24:49 +0000 (0:00:04.287)       0:02:39.810 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Thursday 26 March 2026  04:24:50 +0000 (0:00:01.359)       0:02:41.170 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Thursday 26 March 2026  04:24:52 +0000 (0:00:01.639)       0:02:42.809 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Thursday 26 March 2026  04:24:54 +0000 (0:00:01.553)       0:02:44.363 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Thursday 26 March 2026  04:24:55 +0000 (0:00:01.704)       0:02:46.067 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Thursday 26 March 2026  04:24:57 +0000 (0:00:01.488)       0:02:47.556 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Thursday 26 March 2026  04:24:58 +0000 (0:00:01.376)       0:02:48.932 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Thursday 26 March 2026  04:25:00 +0000 (0:00:01.437)       0:02:50.370 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Thursday 26 March 2026  04:25:01 +0000 (0:00:01.714)       0:02:52.085 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Thursday 26 March 2026  04:25:03 +0000 (0:00:02.180)       0:02:54.265 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Thursday 26 March 2026  04:25:06 +0000 (0:00:02.250)       0:02:56.516 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Thursday 26 March 2026  04:25:14 +0000 (0:00:07.927)       0:03:04.443 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Thursday 26 March 2026  04:25:16 +0000 (0:00:02.221)       0:03:06.664 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Thursday 26 March 2026  04:25:17 +0000 (0:00:01.434)       0:03:08.099 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Thursday 26 March 2026  04:25:19 +0000 (0:00:01.540)       0:03:09.640 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Thursday 26 March 2026  04:25:21 +0000 (0:00:01.925)       0:03:11.565 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Thursday 26 March 2026  04:25:22 +0000 (0:00:01.537)       0:03:13.103 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Thursday 26 March 2026  04:25:25 +0000 (0:00:02.937)       0:03:16.040 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Thursday 26 March 2026  04:25:27 +0000 (0:00:01.787)       0:03:17.828 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Thursday 26 March 2026  04:25:28 +0000 (0:00:01.448)       0:03:19.277 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Thursday 26 March 2026  04:25:30 +0000 (0:00:01.867)       0:03:21.144 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Thursday 26 March 2026  04:25:31 +0000 (0:00:01.154)       0:03:22.299 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Thursday 26 March 2026  04:25:33 +0000 (0:00:01.486)       0:03:23.785 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Thursday 26 March 2026  04:25:35 +0000 (0:00:01.834)       0:03:25.620 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Thursday 26 March 2026  04:25:37 +0000 (0:00:02.573)       0:03:28.194 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Thursday 26 March 2026  04:25:39 +0000 (0:00:01.514)       0:03:29.709 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Thursday 26 March 2026  04:25:40 +0000 (0:00:01.433)       0:03:31.143 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Thursday 26 March 2026  04:25:42 +0000 (0:00:01.654)       0:03:32.798 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Thursday 26 March 2026  04:25:44 +0000 (0:00:02.418)       0:03:35.217 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Thursday 26 March 2026  04:25:46 +0000 (0:00:01.700)       0:03:36.917 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Thursday 26 March 2026  04:25:47 +0000 (0:00:01.410)       0:03:38.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Thursday 26 March 2026  04:25:49 +0000 (0:00:01.715)       0:03:40.044 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Thursday 26 March 2026  04:25:51 +0000 (0:00:01.847)       0:03:41.892 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Thursday 26 March 2026  04:25:53 +0000 (0:00:01.905)       0:03:43.797 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Thursday 26 March 2026  04:25:55 +0000 (0:00:01.865)       0:03:45.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Thursday 26 March 2026  04:25:56 +0000 (0:00:01.151)       0:03:46.815 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Thursday 26 March 2026  04:25:58 +0000 (0:00:01.972)       0:03:48.788 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Thursday 26 March 2026  04:26:00 +0000 (0:00:02.341)       0:03:51.130 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Thursday 26 March 2026  04:26:01 +0000 (0:00:01.175)       0:03:52.306 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Thursday 26 March 2026  04:26:03 +0000 (0:00:01.170)       0:03:53.476 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"Installed Cilium version: 1.18.2, Target Cilium version: v1.18.2, Update needed: False\\n\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Thursday 26 March 2026  04:26:04 +0000 (0:00:01.149)       0:03:54.626 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Thursday 26 March 2026  04:26:05 +0000 (0:00:01.155)       0:03:55.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Thursday 26 March 2026  04:26:10 +0000 (0:00:05.467)       0:04:01.250 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Thursday 26 March 2026  04:26:13 +0000 (0:00:02.383)       0:04:03.634 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Thursday 26 March 2026  04:26:15 +0000 (0:00:02.620)       0:04:06.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Thursday 26 March 2026  04:26:19 +0000 (0:00:04.090)       0:04:10.345 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Thursday 26 March 2026  04:26:21 +0000 (0:00:01.145)       0:04:11.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Thursday 26 March 2026  04:26:24 +0000 (0:00:02.908)       0:04:14.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Thursday 26 March 2026  04:26:25 +0000 (0:00:01.461)       0:04:15.861 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Thursday 26 March 2026  04:26:27 +0000 (0:00:02.085)       0:04:17.946 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Thursday 26 March 2026  04:26:28 +0000 (0:00:01.326)       0:04:19.273 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Thursday 26 March 2026  04:26:30 +0000 (0:00:01.561)       0:04:20.835 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Thursday 26 March 2026  04:26:35 +0000 (0:00:05.003)       0:04:25.839 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Thursday 26 March 2026  04:26:37 +0000 (0:00:01.875)       0:04:27.715 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Thursday 26 March 2026  04:26:46 +0000 (0:00:09.159)       0:04:36.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Thursday 26 March 2026  04:26:48 +0000 (0:00:01.861)       0:04:38.736 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=53  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:26:50 +0000 (0:00:02.479)       0:04:41.215 ******** ",
                                "=============================================================================== ",
                                "k3s_server : Enable and check K3s service ------------------------------ 27.06s",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 23.37s",
                                "Manage labels ----------------------------------------------------------- 9.16s",
                                "k3s_agent : Manage k3s service ------------------------------------------ 7.93s",
                                "k3s_server_post : Wait for Cilium resources ----------------------------- 5.47s",
                                "k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites --- 5.41s",
                                "k9s : Install k9s packages ---------------------------------------------- 5.00s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 4.29s",
                                "k3s_server_post : Apply BGP manifests ----------------------------------- 4.09s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 3.15s",
                                "k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries --- 3.01s",
                                "Make kubeconfig available for use inside the manager service ------------ 2.94s",
                                "k3s_server_post : Test for BGP config resources ------------------------- 2.91s",
                                "k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 2.86s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 2.69s",
                                "k3s_server_post : Copy BGP manifests to first master -------------------- 2.62s",
                                "kubectl : Install apt-transport-https package --------------------------- 2.57s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.53s",
                                "k3s_server : Copy vip manifest to first master -------------------------- 2.52s",
                                "Manage taints ----------------------------------------------------------- 2.48s",
                                "+ [[ false == \\f\\a\\l\\s\\e ]]",
                                "+ sh -c /opt/configuration/scripts/upgrade/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ osism apply openstackclient",
                                "\u001b[32m2026-03-26 04:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e1a66714-e3a4-43c7-84c4-5634d0d755e8 (openstackclient) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e1a66714-e3a4-43c7-84c4-5634d0d755e8 (openstackclient) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x7b96ee349be0>): Expecting value: line 2 column 1 (char 1)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x7b96ee349be0>): 'NoneType' object is not subscriptable\u001b[0m",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Thursday 26 March 2026  04:27:10 +0000 (0:00:01.545)       0:00:01.545 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Thursday 26 March 2026  04:27:10 +0000 (0:00:00.865)       0:00:02.410 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Thursday 26 March 2026  04:27:12 +0000 (0:00:01.489)       0:00:03.900 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Thursday 26 March 2026  04:27:13 +0000 (0:00:01.268)       0:00:05.168 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Thursday 26 March 2026  04:27:14 +0000 (0:00:01.125)       0:00:06.293 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Thursday 26 March 2026  04:27:15 +0000 (0:00:00.931)       0:00:07.225 ******** ",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m",
                                "\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x7b96ee349be0>): 'NoneType' object is not subscriptable\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Thursday 26 March 2026  04:27:16 +0000 (0:00:00.713)       0:00:07.938 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Thursday 26 March 2026  04:27:26 +0000 (0:00:10.507)       0:00:18.446 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Thursday 26 March 2026  04:27:28 +0000 (0:00:01.309)       0:00:19.755 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Thursday 26 March 2026  04:27:28 +0000 (0:00:00.616)       0:00:20.371 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:27:30 +0000 (0:00:01.156)       0:00:21.528 ******** ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Restart openstackclient service ------- 10.51s",
                                "osism.services.openstackclient : Create required directories ------------ 1.49s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 1.31s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 1.27s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 1.16s",
                                "osism.services.openstackclient : Manage openstackclient service --------- 1.13s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 0.93s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.87s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 0.71s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.62s",
                                "+ osism apply -a upgrade common",
                                "\u001b[32m2026-03-26 04:27:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d9da0468-b2fe-479e-8776-94e7f6347ff6 (common) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:27:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d9da0468-b2fe-479e-8776-94e7f6347ff6 (common) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:27:39 +0000 (0:00:02.502)       0:00:02.502 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/upgrade.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  04:27:43 +0000 (0:00:03.336)       0:00:05.838 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:27:46 +0000 (0:00:03.585)       0:00:09.424 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  04:27:49 +0000 (0:00:02.808)       0:00:12.232 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:27:53 +0000 (0:00:04.313)       0:00:16.546 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  04:27:56 +0000 (0:00:02.875)       0:00:19.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensure /var/log/journal exists on EL10 systems] *****************",
                                "Thursday 26 March 2026  04:27:59 +0000 (0:00:03.287)       0:00:22.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Thursday 26 March 2026  04:28:02 +0000 (0:00:02.349)       0:00:25.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Thursday 26 March 2026  04:28:04 +0000 (0:00:02.157)       0:00:27.215 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over kolla.target] **************************************",
                                "Thursday 26 March 2026  04:28:06 +0000 (0:00:02.257)       0:00:29.473 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  04:28:09 +0000 (0:00:02.933)       0:00:32.406 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Thursday 26 March 2026  04:28:14 +0000 (0:00:04.925)       0:00:37.331 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Thursday 26 March 2026  04:28:16 +0000 (0:00:02.395)       0:00:39.727 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Thursday 26 March 2026  04:28:18 +0000 (0:00:01.857)       0:00:41.585 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Thursday 26 March 2026  04:28:20 +0000 (0:00:01.853)       0:00:43.438 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Thursday 26 March 2026  04:28:22 +0000 (0:00:01.835)       0:00:45.273 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Thursday 26 March 2026  04:28:26 +0000 (0:00:04.267)       0:00:49.541 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Thursday 26 March 2026  04:28:29 +0000 (0:00:03.102)       0:00:52.643 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Thursday 26 March 2026  04:28:32 +0000 (0:00:02.866)       0:00:55.509 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Thursday 26 March 2026  04:28:35 +0000 (0:00:02.813)       0:00:58.323 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Thursday 26 March 2026  04:28:38 +0000 (0:00:03.118)       0:01:01.441 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Check containers] ********************",
                                "Thursday 26 March 2026  04:28:41 +0000 (0:00:03.309)       0:01:04.751 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:28:46 +0000 (0:00:04.481)       0:01:09.233 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:28:48 +0000 (0:00:02.103)       0:01:11.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:51 +0000 (0:00:03.054)       0:01:14.392 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:52 +0000 (0:00:00.446)       0:01:14.838 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:52 +0000 (0:00:00.441)       0:01:15.280 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:52 +0000 (0:00:00.442)       0:01:15.722 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:53 +0000 (0:00:00.461)       0:01:16.183 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:54 +0000 (0:00:00.715)       0:01:16.899 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:28:54 +0000 (0:00:00.443)       0:01:17.342 ******** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Thursday 26 March 2026  04:28:55 +0000 (0:00:00.829)       0:01:18.172 ******** ",
                                "\u001b[0;31mfatal: [testbed-manager]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_q4x7yxbs/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_dnvx1h1y/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-3]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_zh6b6vk4/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_bqyc2qzf/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-4]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_c5169jl2/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[32m2026-03-26 04:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62611976-a838-4c25-9ed0-48c8dae1f80c (common) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62611976-a838-4c25-9ed0-48c8dae1f80c (common) has been started and output is visible here.\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_tymil31q/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-5]: FAILED! => {\"changed\": true, \"msg\": \"'Traceback (most recent call last):\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 275, in _raise_for_status\\\\n    response.raise_for_status()\\\\n  File \\\"/usr/lib/python3/dist-packages/requests/models.py\\\", line 1021, in raise_for_status\\\\n    raise HTTPError(http_error_msg, response=self)\\\\nrequests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd\\\\n\\\\nThe above exception was the direct cause of the following exception:\\\\n\\\\nTraceback (most recent call last):\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/modules/kolla_container.py\\\", line 421, in main\\\\n    result = bool(getattr(cw, module.params.get(\\\\'action\\\\'))())\\\\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 361, in recreate_or_restart_container\\\\n    self.pull_image()\\\\n  File \\\"/tmp/ansible_kolla_container_payload_jm1xy0ab/ansible_kolla_container_payload.zip/ansible/module_utils/kolla_docker_worker.py\\\", line 202, in pull_image\\\\n    json.loads(line.strip().decode(\\\\'utf-8\\\\')) for line in self.dc.pull(\\\\n                                                         ^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/image.py\\\", line 429, in pull\\\\n    self._raise_for_status(response)\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/api/client.py\\\", line 277, in _raise_for_status\\\\n    raise create_api_error_from_http_exception(e) from e\\\\n          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\\\\n  File \\\"/usr/lib/python3/dist-packages/docker/errors.py\\\", line 39, in create_api_error_from_http_exception\\\\n    raise cls(e, response=response, explanation=explanation) from e\\\\ndocker.errors.APIError: 500 Server Error for http+docker://localhost/v1.47/images/create?tag=5.0.8.20251208&fromImage=registry.osism.tech%2Fkolla%2Frelease%2Ffluentd: Internal Server Error (\\\"unknown: artifact kolla/release/fluentd:5.0.8.20251208 not found\\\")\\\\n'\"}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;31mtestbed-manager\u001b[0m            : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-0\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-3\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-4\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;31mtestbed-node-5\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:29:01 +0000 (0:00:06.225)       0:01:24.397 ******** ",
                                "=============================================================================== ",
                                "common : Restart fluentd container -------------------------------------- 6.23s",
                                "common : Copying over config.json files for services -------------------- 4.93s",
                                "service-check-containers : common | Check containers -------------------- 4.48s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 4.31s",
                                "common : Copying over fluentd.conf -------------------------------------- 4.27s",
                                "common : Flush handlers ------------------------------------------------- 3.78s",
                                "common : Ensuring config directories exist ------------------------------ 3.59s",
                                "common : include_tasks -------------------------------------------------- 3.34s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 3.31s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 3.29s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 3.12s",
                                "common : Copying over cron logrotate config file ------------------------ 3.10s",
                                "service-check-containers : Include tasks -------------------------------- 3.05s",
                                "common : Copying over kolla.target -------------------------------------- 2.93s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 2.88s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.87s",
                                "common : Ensuring config directories have correct owner and permission --- 2.81s",
                                "common : include_tasks -------------------------------------------------- 2.81s",
                                "common : Find custom fluentd input config files ------------------------- 2.40s",
                                "common : Ensure /var/log/journal exists on EL10 systems ----------------- 2.35s",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x774c1743cad0>): Expecting value: line 2 column 1 (char 1)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x774c1743cad0>): 'NoneType' object is not subscriptable\u001b[0m",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:29:07 +0000 (0:00:01.579)       0:00:01.581 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/upgrade.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  04:29:10 +0000 (0:00:02.455)       0:00:04.036 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:29:12 +0000 (0:00:02.257)       0:00:06.294 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Thursday 26 March 2026  04:29:14 +0000 (0:00:02.087)       0:00:08.381 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:29:18 +0000 (0:00:03.314)       0:00:11.696 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Thursday 26 March 2026  04:29:19 +0000 (0:00:01.538)       0:00:13.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensure /var/log/journal exists on EL10 systems] *****************",
                                "Thursday 26 March 2026  04:29:21 +0000 (0:00:02.256)       0:00:15.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Thursday 26 March 2026  04:29:22 +0000 (0:00:01.071)       0:00:16.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Thursday 26 March 2026  04:29:23 +0000 (0:00:00.929)       0:00:17.492 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over kolla.target] **************************************",
                                "Thursday 26 March 2026  04:29:24 +0000 (0:00:00.789)       0:00:18.281 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  04:29:26 +0000 (0:00:01.923)       0:00:20.205 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Thursday 26 March 2026  04:29:30 +0000 (0:00:03.670)       0:00:23.875 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Thursday 26 March 2026  04:29:31 +0000 (0:00:01.454)       0:00:25.330 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Thursday 26 March 2026  04:29:32 +0000 (0:00:00.897)       0:00:26.228 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Thursday 26 March 2026  04:29:33 +0000 (0:00:00.949)       0:00:27.177 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Thursday 26 March 2026  04:29:34 +0000 (0:00:00.935)       0:00:28.112 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Thursday 26 March 2026  04:29:37 +0000 (0:00:02.815)       0:00:30.927 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Thursday 26 March 2026  04:29:39 +0000 (0:00:02.467)       0:00:33.395 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Thursday 26 March 2026  04:29:41 +0000 (0:00:01.984)       0:00:35.380 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Thursday 26 March 2026  04:29:43 +0000 (0:00:02.110)       0:00:37.490 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Thursday 26 March 2026  04:29:46 +0000 (0:00:02.170)       0:00:39.661 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Check containers] ********************",
                                "Thursday 26 March 2026  04:29:48 +0000 (0:00:02.193)       0:00:41.854 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:29:51 +0000 (0:00:03.167)       0:00:45.021 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:29:52 +0000 (0:00:01.045)       0:00:46.067 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m",
                                "\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x774c1743cad0>): 'NoneType' object is not subscriptable\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/fluentd:5.0.8.20251208', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/kolla-toolbox:20.3.1.20251208', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cron:3.0.20251208', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:54 +0000 (0:00:02.275)       0:00:48.342 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:54 +0000 (0:00:00.090)       0:00:48.433 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:54 +0000 (0:00:00.075)       0:00:48.508 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:54 +0000 (0:00:00.072)       0:00:48.581 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:55 +0000 (0:00:00.072)       0:00:48.653 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:55 +0000 (0:00:00.369)       0:00:49.022 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  04:29:55 +0000 (0:00:00.080)       0:00:49.103 ******** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Thursday 26 March 2026  04:29:55 +0000 (0:00:00.109)       0:00:49.212 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Thursday 26 March 2026  04:30:33 +0000 (0:00:37.548)       0:01:26.760 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Thursday 26 March 2026  04:31:12 +0000 (0:00:39.056)       0:02:05.817 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Thursday 26 March 2026  04:31:14 +0000 (0:00:01.973)       0:02:07.790 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:31:26 +0000 (0:00:11.972)       0:02:19.762 ******** ",
                                "=============================================================================== ",
                                "common : Restart kolla-toolbox container ------------------------------- 39.06s",
                                "common : Restart fluentd container ------------------------------------- 37.55s",
                                "common : Restart cron container ---------------------------------------- 11.97s",
                                "common : Copying over config.json files for services -------------------- 3.67s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 3.31s",
                                "service-check-containers : common | Check containers -------------------- 3.17s",
                                "common : Copying over fluentd.conf -------------------------------------- 2.82s",
                                "common : Copying over cron logrotate config file ------------------------ 2.47s",
                                "common : include_tasks -------------------------------------------------- 2.46s",
                                "service-check-containers : Include tasks -------------------------------- 2.28s",
                                "common : Ensuring config directories exist ------------------------------ 2.26s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 2.26s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.19s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.17s",
                                "common : Ensuring config directories have correct owner and permission --- 2.11s",
                                "common : include_tasks -------------------------------------------------- 2.09s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 1.99s",
                                "common : Initializing toolbox container using normal user --------------- 1.97s",
                                "common : Copying over kolla.target -------------------------------------- 1.92s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 1.54s",
                                "+ osism apply -a upgrade loadbalancer",
                                "\u001b[32m2026-03-26 04:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5147d542-5a4e-40bb-a5dc-01c2c9e8a857 (loadbalancer) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5147d542-5a4e-40bb-a5dc-01c2c9e8a857 (loadbalancer) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:31:36 +0000 (0:00:02.439)       0:00:02.439 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:31:38 +0000 (0:00:02.336)       0:00:04.775 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Thursday 26 March 2026  04:31:42 +0000 (0:00:03.757)       0:00:08.533 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Stop and remove containers for haproxy exporter containers] ***",
                                "Thursday 26 March 2026  04:31:44 +0000 (0:00:02.088)       0:00:10.621 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing config for haproxy exporter] *********************",
                                "Thursday 26 March 2026  04:31:46 +0000 (0:00:02.060)       0:00:12.682 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Thursday 26 March 2026  04:31:48 +0000 (0:00:02.137)       0:00:14.820 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Thursday 26 March 2026  04:31:50 +0000 (0:00:01.839)       0:00:16.660 ******** ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Thursday 26 March 2026  04:31:52 +0000 (0:00:01.921)       0:00:18.581 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Thursday 26 March 2026  04:31:53 +0000 (0:00:01.697)       0:00:20.279 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Thursday 26 March 2026  04:31:58 +0000 (0:00:04.044)       0:00:24.323 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Thursday 26 March 2026  04:32:00 +0000 (0:00:02.183)       0:00:26.507 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Thursday 26 March 2026  04:32:02 +0000 (0:00:02.350)       0:00:28.858 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Thursday 26 March 2026  04:32:04 +0000 (0:00:01.899)       0:00:30.758 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Thursday 26 March 2026  04:32:07 +0000 (0:00:02.672)       0:00:33.430 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Thursday 26 March 2026  04:32:09 +0000 (0:00:01.982)       0:00:35.412 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Thursday 26 March 2026  04:32:11 +0000 (0:00:02.727)       0:00:38.141 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Thursday 26 March 2026  04:32:14 +0000 (0:00:02.315)       0:00:40.456 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Thursday 26 March 2026  04:32:16 +0000 (0:00:02.341)       0:00:42.798 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Thursday 26 March 2026  04:32:18 +0000 (0:00:01.712)       0:00:44.511 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy-ssh:9.6.20251208', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689', '__omit_place_holder__963cad7ce6778b422210f936b7a59b3fd90ba689'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Thursday 26 March 2026  04:32:23 +0000 (0:00:05.084)       0:00:49.596 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Thursday 26 March 2026  04:32:28 +0000 (0:00:04.771)       0:00:54.367 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Thursday 26 March 2026  04:32:30 +0000 (0:00:02.726)       0:00:57.094 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Thursday 26 March 2026  04:32:35 +0000 (0:00:04.402)       0:01:01.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Thursday 26 March 2026  04:32:37 +0000 (0:00:01.896)       0:01:03.393 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Thursday 26 March 2026  04:32:40 +0000 (0:00:02.967)       0:01:06.361 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Thursday 26 March 2026  04:32:42 +0000 (0:00:02.798)       0:01:09.159 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Thursday 26 March 2026  04:32:44 +0000 (0:00:01.863)       0:01:11.023 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Thursday 26 March 2026  04:32:47 +0000 (0:00:02.690)       0:01:13.713 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql-cert.pem] ***************************",
                                "Thursday 26 March 2026  04:32:49 +0000 (0:00:02.590)       0:01:16.303 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql-key.pem] ****************************",
                                "Thursday 26 March 2026  04:32:51 +0000 (0:00:01.409)       0:01:17.713 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  04:32:53 +0000 (0:00:01.933)       0:01:19.647 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:32:57 +0000 (0:00:04.123)       0:01:23.771 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  04:32:59 +0000 (0:00:01.634)       0:01:25.405 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Thursday 26 March 2026  04:33:01 +0000 (0:00:02.100)       0:01:27.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Thursday 26 March 2026  04:33:03 +0000 (0:00:02.576)       0:01:30.082 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Thursday 26 March 2026  04:33:06 +0000 (0:00:02.495)       0:01:32.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Check containers] **************",
                                "Thursday 26 March 2026  04:33:08 +0000 (0:00:02.658)       0:01:35.236 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:33:13 +0000 (0:00:04.113)       0:01:39.350 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:33:14 +0000 (0:00:01.395)       0:01:40.745 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Thursday 26 March 2026  04:33:16 +0000 (0:00:02.265)       0:01:43.010 ******** ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Thursday 26 March 2026  04:33:18 +0000 (0:00:02.022)       0:01:45.032 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Thursday 26 March 2026  04:33:23 +0000 (0:00:04.748)       0:01:49.781 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-api:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-evaluator:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-listener:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/aodh-notifier:20.0.0.20251208', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Thursday 26 March 2026  04:33:25 +0000 (0:00:01.719)       0:01:51.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Thursday 26 March 2026  04:33:27 +0000 (0:00:02.425)       0:01:53.927 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Thursday 26 March 2026  04:33:29 +0000 (0:00:02.386)       0:01:56.313 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Thursday 26 March 2026  04:33:32 +0000 (0:00:02.961)       0:01:59.275 ******** ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Thursday 26 March 2026  04:33:34 +0000 (0:00:01.605)       0:02:00.881 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Thursday 26 March 2026  04:33:39 +0000 (0:00:04.523)       0:02:05.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-api:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-keystone-listener:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/barbican-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Thursday 26 March 2026  04:33:40 +0000 (0:00:01.626)       0:02:07.030 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Thursday 26 March 2026  04:33:42 +0000 (0:00:01.806)       0:02:08.837 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  04:33:44 +0000 (0:00:02.341)       0:02:11.178 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Thursday 26 March 2026  04:33:47 +0000 (0:00:02.844)       0:02:14.023 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Thursday 26 March 2026  04:33:49 +0000 (0:00:01.352)       0:02:15.375 ******** ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Thursday 26 March 2026  04:33:50 +0000 (0:00:01.728)       0:02:17.104 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Thursday 26 March 2026  04:33:54 +0000 (0:00:03.538)       0:02:20.643 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Thursday 26 March 2026  04:33:56 +0000 (0:00:02.633)       0:02:23.276 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Thursday 26 March 2026  04:33:59 +0000 (0:00:02.867)       0:02:26.143 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  04:34:01 +0000 (0:00:01.591)       0:02:27.735 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Thursday 26 March 2026  04:34:03 +0000 (0:00:02.474)       0:02:30.209 ******** ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Thursday 26 March 2026  04:34:05 +0000 (0:00:01.805)       0:02:32.015 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Thursday 26 March 2026  04:34:10 +0000 (0:00:04.835)       0:02:36.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-api:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-scheduler:26.2.1.20251208', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-volume:26.2.1.20251208', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/cinder-backup:26.2.1.20251208', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Thursday 26 March 2026  04:34:12 +0000 (0:00:01.941)       0:02:38.792 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Thursday 26 March 2026  04:34:14 +0000 (0:00:02.014)       0:02:40.806 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Thursday 26 March 2026  04:34:16 +0000 (0:00:02.237)       0:02:43.043 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Thursday 26 March 2026  04:34:19 +0000 (0:00:02.849)       0:02:45.893 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Thursday 26 March 2026  04:34:21 +0000 (0:00:01.567)       0:02:47.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Thursday 26 March 2026  04:34:22 +0000 (0:00:01.398)       0:02:48.859 ******** ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Thursday 26 March 2026  04:34:24 +0000 (0:00:01.783)       0:02:50.643 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Thursday 26 March 2026  04:34:29 +0000 (0:00:04.864)       0:02:55.508 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-api:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-backend-bind9:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-central:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-mdns:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-producer:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-worker:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/designate-sink:20.0.1.20251208', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Thursday 26 March 2026  04:34:31 +0000 (0:00:01.914)       0:02:57.422 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:34:33 +0000 (0:00:02.174)       0:02:59.597 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:34:35 +0000 (0:00:02.304)       0:03:01.901 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Thursday 26 March 2026  04:34:38 +0000 (0:00:02.923)       0:03:04.825 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Thursday 26 March 2026  04:34:39 +0000 (0:00:01.332)       0:03:06.157 ******** ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Thursday 26 March 2026  04:34:41 +0000 (0:00:01.819)       0:03:07.977 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Thursday 26 March 2026  04:34:47 +0000 (0:00:05.708)       0:03:13.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/glance-api:30.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/glance-tls-proxy:30.0.1.20251208', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Thursday 26 March 2026  04:34:51 +0000 (0:00:04.576)       0:03:18.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Thursday 26 March 2026  04:34:56 +0000 (0:00:04.653)       0:03:22.916 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Thursday 26 March 2026  04:34:58 +0000 (0:00:02.295)       0:03:25.211 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Thursday 26 March 2026  04:35:01 +0000 (0:00:02.964)       0:03:28.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Thursday 26 March 2026  04:35:03 +0000 (0:00:01.641)       0:03:29.817 ******** ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Thursday 26 March 2026  04:35:05 +0000 (0:00:01.735)       0:03:31.553 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Thursday 26 March 2026  04:35:09 +0000 (0:00:04.170)       0:03:35.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/grafana:12.3.0.20251208', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Thursday 26 March 2026  04:35:10 +0000 (0:00:01.512)       0:03:37.235 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:35:12 +0000 (0:00:01.381)       0:03:38.616 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:35:14 +0000 (0:00:02.275)       0:03:40.892 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Thursday 26 March 2026  04:35:17 +0000 (0:00:02.862)       0:03:43.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Thursday 26 March 2026  04:35:18 +0000 (0:00:01.425)       0:03:45.180 ******** ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Thursday 26 March 2026  04:35:20 +0000 (0:00:01.779)       0:03:46.960 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Thursday 26 March 2026  04:35:25 +0000 (0:00:04.915)       0:03:51.875 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/horizon:25.3.2.20251208', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Thursday 26 March 2026  04:35:27 +0000 (0:00:01.701)       0:03:53.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:35:29 +0000 (0:00:02.029)       0:03:55.607 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:35:31 +0000 (0:00:02.253)       0:03:57.860 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Thursday 26 March 2026  04:35:34 +0000 (0:00:02.893)       0:04:00.754 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Thursday 26 March 2026  04:35:35 +0000 (0:00:01.395)       0:04:02.150 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Thursday 26 March 2026  04:35:37 +0000 (0:00:01.364)       0:04:03.514 ******** ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Thursday 26 March 2026  04:35:39 +0000 (0:00:02.025)       0:04:05.540 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Thursday 26 March 2026  04:35:44 +0000 (0:00:04.854)       0:04:10.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-ssh:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keystone-fernet:27.0.1.20251208', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Thursday 26 March 2026  04:35:46 +0000 (0:00:02.048)       0:04:12.444 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Thursday 26 March 2026  04:35:47 +0000 (0:00:01.707)       0:04:14.151 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  04:35:50 +0000 (0:00:02.283)       0:04:16.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Thursday 26 March 2026  04:35:53 +0000 (0:00:03.304)       0:04:19.739 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Thursday 26 March 2026  04:35:54 +0000 (0:00:01.383)       0:04:21.123 ******** ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Thursday 26 March 2026  04:35:56 +0000 (0:00:01.920)       0:04:23.044 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Thursday 26 March 2026  04:36:01 +0000 (0:00:04.911)       0:04:27.955 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-api:20.0.1.20251208', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/magnum-conductor:20.0.1.20251208', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Thursday 26 March 2026  04:36:03 +0000 (0:00:01.862)       0:04:29.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Thursday 26 March 2026  04:36:05 +0000 (0:00:02.006)       0:04:31.823 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Thursday 26 March 2026  04:36:07 +0000 (0:00:02.371)       0:04:34.195 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Thursday 26 March 2026  04:36:10 +0000 (0:00:02.925)       0:04:37.121 ******** ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Thursday 26 March 2026  04:36:12 +0000 (0:00:02.165)       0:04:39.286 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Thursday 26 March 2026  04:36:17 +0000 (0:00:04.923)       0:04:44.209 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-api:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-scheduler:20.0.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-share:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/2025.1/manila-data:20.0.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Thursday 26 March 2026  04:36:19 +0000 (0:00:01.716)       0:04:45.926 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Thursday 26 March 2026  04:36:21 +0000 (0:00:02.003)       0:04:47.929 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Thursday 26 March 2026  04:36:23 +0000 (0:00:02.254)       0:04:50.183 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Thursday 26 March 2026  04:36:26 +0000 (0:00:02.974)       0:04:53.158 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Thursday 26 March 2026  04:36:29 +0000 (0:00:02.622)       0:04:55.780 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Thursday 26 March 2026  04:36:33 +0000 (0:00:03.966)       0:04:59.747 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Thursday 26 March 2026  04:36:36 +0000 (0:00:03.513)       0:05:03.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-clustercheck:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Thursday 26 March 2026  04:36:40 +0000 (0:00:03.738)       0:05:06.999 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:36:44 +0000 (0:00:03.918)       0:05:10.918 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:36:47 +0000 (0:00:02.974)       0:05:13.893 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Thursday 26 March 2026  04:36:50 +0000 (0:00:02.727)       0:05:16.621 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Thursday 26 March 2026  04:36:51 +0000 (0:00:01.402)       0:05:18.023 ******** ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Thursday 26 March 2026  04:36:53 +0000 (0:00:02.176)       0:05:20.199 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Thursday 26 March 2026  04:36:56 +0000 (0:00:02.535)       0:05:22.735 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Thursday 26 March 2026  04:36:58 +0000 (0:00:01.745)       0:05:24.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:36:59 +0000 (0:00:01.420)       0:05:25.901 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:37:01 +0000 (0:00:01.454)       0:05:27.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Thursday 26 March 2026  04:37:03 +0000 (0:00:02.362)       0:05:29.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Thursday 26 March 2026  04:37:05 +0000 (0:00:01.707)       0:05:31.425 ******** ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Thursday 26 March 2026  04:37:07 +0000 (0:00:02.122)       0:05:33.547 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Thursday 26 March 2026  04:37:13 +0000 (0:00:06.368)       0:05:39.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-server:26.0.3.20251208', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-openvswitch-agent:26.0.3.20251208', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-dhcp-agent:26.0.3.20251208', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/release/2025.1/neutron-l3-agent:26.0.3.20251208', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-sriov-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-mlnx-agent:26.0.3.20251208', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-eswitchd:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metadata-agent:26.0.3.20251208', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-bgp-dragent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-infoblox-ipam-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-metering-agent:26.0.3.20251208', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/2025.1/ironic-neutron-agent:26.0.3.20251208', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/neutron-tls-proxy:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/2025.1/neutron-ovn-agent:26.0.3.20251208', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Thursday 26 March 2026  04:37:16 +0000 (0:00:02.651)       0:05:42.568 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:37:19 +0000 (0:00:02.989)       0:05:45.557 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:37:21 +0000 (0:00:02.255)       0:05:47.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Thursday 26 March 2026  04:37:24 +0000 (0:00:02.986)       0:05:50.800 ******** ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Thursday 26 March 2026  04:37:26 +0000 (0:00:02.470)       0:05:53.271 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Thursday 26 March 2026  04:37:31 +0000 (0:00:04.663)       0:05:57.934 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/2025.1/placement-api:13.0.0.20251208', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Thursday 26 March 2026  04:37:33 +0000 (0:00:01.726)       0:05:59.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:37:35 +0000 (0:00:01.907)       0:06:01.568 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:37:37 +0000 (0:00:02.377)       0:06:03.945 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Thursday 26 March 2026  04:37:40 +0000 (0:00:02.975)       0:06:06.921 ******** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Thursday 26 March 2026  04:37:42 +0000 (0:00:02.309)       0:06:09.231 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Thursday 26 March 2026  04:37:50 +0000 (0:00:07.268)       0:06:16.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-api:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-scheduler:31.2.1.20251208', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/2025.1/nova-super-conductor:31.2.1.20251208', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Thursday 26 March 2026  04:37:52 +0000 (0:00:01.951)       0:06:18.451 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Thursday 26 March 2026  04:37:54 +0000 (0:00:02.577)       0:06:21.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Thursday 26 March 2026  04:37:57 +0000 (0:00:02.333)       0:06:23.362 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Thursday 26 March 2026  04:38:00 +0000 (0:00:03.004)       0:06:26.367 ******** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Thursday 26 March 2026  04:38:02 +0000 (0:00:02.853)       0:06:29.220 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Thursday 26 March 2026  04:38:04 +0000 (0:00:01.765)       0:06:30.986 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Thursday 26 March 2026  04:38:09 +0000 (0:00:04.834)       0:06:35.820 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Thursday 26 March 2026  04:38:11 +0000 (0:00:02.439)       0:06:38.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:38:15 +0000 (0:00:03.286)       0:06:41.547 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:38:19 +0000 (0:00:03.873)       0:06:45.420 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Thursday 26 March 2026  04:38:22 +0000 (0:00:03.493)       0:06:48.914 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Thursday 26 March 2026  04:38:24 +0000 (0:00:01.563)       0:06:50.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Thursday 26 March 2026  04:38:26 +0000 (0:00:02.273)       0:06:52.752 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Thursday 26 March 2026  04:38:29 +0000 (0:00:02.574)       0:06:55.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:38:31 +0000 (0:00:02.312)       0:06:57.639 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:38:35 +0000 (0:00:03.768)       0:07:01.407 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Thursday 26 March 2026  04:38:39 +0000 (0:00:03.955)       0:07:05.363 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Thursday 26 March 2026  04:38:41 +0000 (0:00:02.298)       0:07:07.662 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Thursday 26 March 2026  04:38:43 +0000 (0:00:02.421)       0:07:10.083 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Thursday 26 March 2026  04:38:46 +0000 (0:00:02.444)       0:07:12.527 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Thursday 26 March 2026  04:38:48 +0000 (0:00:02.499)       0:07:15.027 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Thursday 26 March 2026  04:38:52 +0000 (0:00:03.669)       0:07:18.697 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Thursday 26 March 2026  04:38:56 +0000 (0:00:04.374)       0:07:23.071 ******** ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Thursday 26 March 2026  04:38:59 +0000 (0:00:02.593)       0:07:25.664 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Thursday 26 March 2026  04:39:04 +0000 (0:00:04.908)       0:07:30.572 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-api:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-driver-agent:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-health-manager:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-housekeeping:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/octavia-worker:16.0.1.20251208', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Thursday 26 March 2026  04:39:06 +0000 (0:00:02.124)       0:07:32.697 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:39:08 +0000 (0:00:02.116)       0:07:34.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:39:10 +0000 (0:00:02.256)       0:07:37.070 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Thursday 26 March 2026  04:39:13 +0000 (0:00:03.090)       0:07:40.160 ******** ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Thursday 26 March 2026  04:39:16 +0000 (0:00:02.574)       0:07:42.734 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Thursday 26 March 2026  04:39:23 +0000 (0:00:07.544)       0:07:50.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Thursday 26 March 2026  04:39:26 +0000 (0:00:02.320)       0:07:52.600 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Thursday 26 March 2026  04:39:28 +0000 (0:00:01.845)       0:07:54.446 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Thursday 26 March 2026  04:39:29 +0000 (0:00:01.483)       0:07:55.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Thursday 26 March 2026  04:39:31 +0000 (0:00:02.323)       0:07:58.252 ******** ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Thursday 26 March 2026  04:39:34 +0000 (0:00:02.722)       0:08:00.974 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Thursday 26 March 2026  04:39:40 +0000 (0:00:06.034)       0:08:07.009 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-server:3.2.1.20251208', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-node-exporter:1.8.2.20251208', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-mysqld-exporter:0.16.0.20251208', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-memcached-exporter:0.15.0.20251208', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-cadvisor:0.49.2.20251208', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-alertmanager:0.28.1.20251208', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-openstack-exporter:1.7.0.20251208', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-elasticsearch-exporter:1.8.0.20251208', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-blackbox-exporter:0.25.0.20251208', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/prometheus-libvirt-exporter:2.2.0.20251208', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Thursday 26 March 2026  04:39:43 +0000 (0:00:02.577)       0:08:09.587 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Thursday 26 March 2026  04:39:45 +0000 (0:00:01.831)       0:08:11.419 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Thursday 26 March 2026  04:39:47 +0000 (0:00:01.975)       0:08:13.394 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Thursday 26 March 2026  04:39:49 +0000 (0:00:02.205)       0:08:15.599 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Thursday 26 March 2026  04:39:51 +0000 (0:00:02.350)       0:08:17.950 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Thursday 26 March 2026  04:39:55 +0000 (0:00:03.955)       0:08:21.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Thursday 26 March 2026  04:39:57 +0000 (0:00:01.510)       0:08:23.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Thursday 26 March 2026  04:39:58 +0000 (0:00:01.471)       0:08:24.888 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Thursday 26 March 2026  04:40:00 +0000 (0:00:01.974)       0:08:26.862 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Thursday 26 March 2026  04:40:02 +0000 (0:00:02.285)       0:08:29.147 ******** ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Thursday 26 March 2026  04:40:05 +0000 (0:00:02.516)       0:08:31.663 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Thursday 26 March 2026  04:40:13 +0000 (0:00:07.717)       0:08:39.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-apiserver:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/skyline-console:6.0.1.20251208', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Thursday 26 March 2026  04:40:14 +0000 (0:00:01.751)       0:08:41.132 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Thursday 26 March 2026  04:40:17 +0000 (0:00:02.189)       0:08:43.322 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Thursday 26 March 2026  04:40:19 +0000 (0:00:02.377)       0:08:45.700 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Thursday 26 March 2026  04:40:22 +0000 (0:00:03.317)       0:08:49.017 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Thursday 26 March 2026  04:40:24 +0000 (0:00:01.504)       0:08:50.522 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Thursday 26 March 2026  04:40:25 +0000 (0:00:01.460)       0:08:51.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Thursday 26 March 2026  04:40:27 +0000 (0:00:01.827)       0:08:53.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Thursday 26 March 2026  04:40:28 +0000 (0:00:01.355)       0:08:55.166 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : loadbalancer] *********************************************",
                                "Thursday 26 March 2026  04:40:30 +0000 (0:00:01.430)       0:08:56.597 ******** ",
                                "\u001b[0;36mincluded: loadbalancer for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Check containers] **************",
                                "Thursday 26 March 2026  04:40:33 +0000 (0:00:02.818)       0:08:59.416 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:40:37 +0000 (0:00:04.383)       0:09:03.799 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:40:38 +0000 (0:00:01.442)       0:09:05.242 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/haproxy:2.8.15.20251208', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/proxysql:3.0.3.20251208', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/keepalived:2.2.8.20251208', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Thursday 26 March 2026  04:40:41 +0000 (0:00:02.937)       0:09:08.180 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Thursday 26 March 2026  04:40:43 +0000 (0:00:01.803)       0:09:09.984 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Thursday 26 March 2026  04:40:45 +0000 (0:00:01.398)       0:09:11.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Thursday 26 March 2026  04:40:52 +0000 (0:00:07.025)       0:09:18.407 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Thursday 26 March 2026  04:40:59 +0000 (0:00:07.440)       0:09:25.848 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Thursday 26 March 2026  04:41:06 +0000 (0:00:07.139)       0:09:32.988 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Thursday 26 March 2026  04:41:14 +0000 (0:00:07.613)       0:09:40.602 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Thursday 26 March 2026  04:41:18 +0000 (0:00:03.825)       0:09:44.428 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Thursday 26 March 2026  04:41:31 +0000 (0:00:13.028)       0:09:57.456 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Thursday 26 March 2026  04:41:35 +0000 (0:00:04.614)       0:10:02.070 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Thursday 26 March 2026  04:41:43 +0000 (0:00:07.753)       0:10:09.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Thursday 26 March 2026  04:41:50 +0000 (0:00:06.838)       0:10:16.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Thursday 26 March 2026  04:41:57 +0000 (0:00:06.796)       0:10:23.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Thursday 26 March 2026  04:42:04 +0000 (0:00:06.957)       0:10:30.417 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for master haproxy to start] **************",
                                "Thursday 26 March 2026  04:42:11 +0000 (0:00:07.033)       0:10:37.451 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Thursday 26 March 2026  04:42:14 +0000 (0:00:03.645)       0:10:41.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for master proxysql to start] *************",
                                "Thursday 26 March 2026  04:42:27 +0000 (0:00:12.640)       0:10:53.737 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Thursday 26 March 2026  04:42:31 +0000 (0:00:03.624)       0:10:57.362 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Thursday 26 March 2026  04:42:37 +0000 (0:00:06.740)       0:11:04.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Thursday 26 March 2026  04:42:39 +0000 (0:00:01.974)       0:11:06.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=129 \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=128 \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=128 \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:42:42 +0000 (0:00:02.803)       0:11:08.880 ******** ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup proxysql container ------------------------- 13.03s",
                                "loadbalancer : Start master proxysql container ------------------------- 12.64s",
                                "loadbalancer : Start backup keepalived container ------------------------ 7.75s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 7.72s",
                                "loadbalancer : Start backup haproxy container --------------------------- 7.61s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 7.54s",
                                "loadbalancer : Stop backup haproxy container ---------------------------- 7.44s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 7.27s",
                                "loadbalancer : Stop backup proxysql container --------------------------- 7.14s",
                                "loadbalancer : Start master haproxy container --------------------------- 7.03s",
                                "loadbalancer : Stop backup keepalived container ------------------------- 7.03s",
                                "loadbalancer : Stop master keepalived container ------------------------- 6.96s",
                                "loadbalancer : Stop master haproxy container ---------------------------- 6.84s",
                                "loadbalancer : Stop master proxysql container --------------------------- 6.80s",
                                "loadbalancer : Start master keepalived container ------------------------ 6.74s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 6.37s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 6.03s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 5.71s",
                                "loadbalancer : Copying checks for services which are enabled ------------ 5.08s",
                                "haproxy-config : Copying over manila haproxy config --------------------- 4.92s",
                                "+ osism apply -a upgrade opensearch",
                                "\u001b[32m2026-03-26 04:42:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 880f3991-9db2-4938-9bc9-5c310c65b633 (opensearch) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:42:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 880f3991-9db2-4938-9bc9-5c310c65b633 (opensearch) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:42:51 +0000 (0:00:01.527)       0:00:01.527 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:42:53 +0000 (0:00:01.660)       0:00:03.188 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  04:42:55 +0000 (0:00:02.422)       0:00:05.611 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Thursday 26 March 2026  04:42:57 +0000 (0:00:02.344)       0:00:07.955 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Thursday 26 March 2026  04:42:59 +0000 (0:00:02.096)       0:00:10.052 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  04:43:02 +0000 (0:00:02.431)       0:00:12.483 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Thursday 26 March 2026  04:43:04 +0000 (0:00:01.804)       0:00:14.287 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:43:07 +0000 (0:00:03.512)       0:00:17.800 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Thursday 26 March 2026  04:43:09 +0000 (0:00:01.824)       0:00:19.625 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Thursday 26 March 2026  04:43:11 +0000 (0:00:01.823)       0:00:21.449 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Thursday 26 March 2026  04:43:15 +0000 (0:00:03.883)       0:00:25.333 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Thursday 26 March 2026  04:43:18 +0000 (0:00:03.433)       0:00:28.766 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : opensearch | Check containers] ****************",
                                "Thursday 26 March 2026  04:43:21 +0000 (0:00:03.101)       0:00:31.867 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : opensearch | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:43:25 +0000 (0:00:03.627)       0:00:35.495 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:43:26 +0000 (0:00:01.361)       0:00:36.856 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch:2.19.4.20251208', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/2025.1/opensearch-dashboards:2.19.4.20251208', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  04:43:28 +0000 (0:00:02.107)       0:00:38.964 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  04:43:30 +0000 (0:00:01.662)       0:00:40.627 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  04:43:30 +0000 (0:00:00.475)       0:00:41.103 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Thursday 26 March 2026  04:43:31 +0000 (0:00:00.445)       0:00:41.549 ******** ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Thursday 26 March 2026  04:43:32 +0000 (0:00:00.853)       0:00:42.402 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Thursday 26 March 2026  04:43:35 +0000 (0:00:03.592)       0:00:45.994 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Thursday 26 March 2026  04:43:40 +0000 (0:00:04.453)       0:00:50.448 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Thursday 26 March 2026  04:44:49 +0000 (0:01:09.534)       0:01:59.982 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Thursday 26 March 2026  04:46:20 +0000 (0:01:30.495)       0:03:30.478 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Thursday 26 March 2026  04:46:22 +0000 (0:00:01.688)       0:03:32.166 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Thursday 26 March 2026  04:46:25 +0000 (0:00:03.316)       0:03:35.483 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Thursday 26 March 2026  04:46:28 +0000 (0:00:03.285)       0:03:38.768 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Thursday 26 March 2026  04:46:29 +0000 (0:00:01.249)       0:03:40.017 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:46:31 +0000 (0:00:01.996)       0:03:42.014 ******** ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 90.50s",
                                "opensearch : Restart opensearch container ------------------------------ 69.53s",
                                "opensearch : Perform a flush -------------------------------------------- 4.45s",
                                "opensearch : Copying over config.json files for services ---------------- 3.88s",
                                "service-check-containers : opensearch | Check containers ---------------- 3.63s",
                                "opensearch : Disable shard allocation ----------------------------------- 3.59s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 3.51s",
                                "opensearch : Copying over opensearch service config file ---------------- 3.43s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 3.32s",
                                "opensearch : Check if a log retention policy exists --------------------- 3.29s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 3.10s",
                                "opensearch : Ensuring config directories exist -------------------------- 2.43s",
                                "Group hosts based on enabled services ----------------------------------- 2.42s",
                                "opensearch : include_tasks ---------------------------------------------- 2.34s",
                                "service-check-containers : Include tasks -------------------------------- 2.11s",
                                "opensearch : Setting sysctl values -------------------------------------- 2.10s",
                                "opensearch : Apply retention policy to existing indices ----------------- 2.00s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.82s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 1.82s",
                                "opensearch : include_tasks ---------------------------------------------- 1.80s",
                                "+ osism apply -a upgrade memcached",
                                "\u001b[32m2026-03-26 04:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08a40978-d999-48b6-a67e-60a08e1bdf47 (memcached) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08a40978-d999-48b6-a67e-60a08e1bdf47 (memcached) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x758c9b818c20>): Expecting value: line 2 column 1 (char 1)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x758c9b818c20>): 'NoneType' object is not subscriptable\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:46:39 +0000 (0:00:01.066)       0:00:01.066 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:46:40 +0000 (0:00:00.952)       0:00:02.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Thursday 26 March 2026  04:46:41 +0000 (0:00:00.878)       0:00:02.897 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Thursday 26 March 2026  04:46:43 +0000 (0:00:01.599)       0:00:04.496 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Thursday 26 March 2026  04:46:44 +0000 (0:00:00.931)       0:00:05.428 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [service-check-containers : memcached | Check containers] *****************",
                                "Thursday 26 March 2026  04:46:46 +0000 (0:00:01.886)       0:00:07.315 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : memcached | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:46:47 +0000 (0:00:01.247)       0:00:08.562 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:46:47 +0000 (0:00:00.353)       0:00:08.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m",
                                "\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x758c9b818c20>): 'NoneType' object is not subscriptable\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/2025.1/memcached:1.6.24.20251208', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Thursday 26 March 2026  04:46:49 +0000 (0:00:01.342)       0:00:10.259 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:46:59 +0000 (0:00:10.074)       0:00:20.334 ******** ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container -------------------------------- 10.08s",
                                "memcached : Copying over config.json files for services ----------------- 1.89s",
                                "memcached : include_tasks ----------------------------------------------- 1.60s",
                                "service-check-containers : Include tasks -------------------------------- 1.34s",
                                "service-check-containers : memcached | Check containers ----------------- 1.25s",
                                "Group hosts based on Kolla action --------------------------------------- 0.95s",
                                "memcached : Ensuring config directories exist --------------------------- 0.93s",
                                "Group hosts based on enabled services ----------------------------------- 0.88s",
                                "service-check-containers : memcached | Notify handlers to restart containers --- 0.35s",
                                "+ osism apply -a upgrade redis",
                                "\u001b[32m2026-03-26 04:47:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 103a9efa-0088-4a4c-9ba5-5bed8cc516cc (redis) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:47:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 103a9efa-0088-4a4c-9ba5-5bed8cc516cc (redis) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:47:07 +0000 (0:00:01.360)       0:00:01.360 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:47:09 +0000 (0:00:02.088)       0:00:03.449 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Thursday 26 March 2026  04:47:12 +0000 (0:00:02.799)       0:00:06.248 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Thursday 26 March 2026  04:47:14 +0000 (0:00:01.728)       0:00:07.977 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Thursday 26 March 2026  04:47:16 +0000 (0:00:02.252)       0:00:10.229 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Thursday 26 March 2026  04:47:19 +0000 (0:00:03.169)       0:00:13.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : redis | Check containers] *********************",
                                "Thursday 26 March 2026  04:47:23 +0000 (0:00:03.891)       0:00:17.291 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : redis | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:47:26 +0000 (0:00:03.135)       0:00:20.426 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:47:28 +0000 (0:00:01.612)       0:00:22.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis:7.0.15.20251208', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/redis-sentinel:7.0.15.20251208', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  04:47:30 +0000 (0:00:01.925)       0:00:23.964 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  04:47:30 +0000 (0:00:00.466)       0:00:24.430 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Thursday 26 March 2026  04:47:31 +0000 (0:00:00.450)       0:00:24.881 ******** ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Thursday 26 March 2026  04:47:32 +0000 (0:00:00.802)       0:00:25.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Thursday 26 March 2026  04:47:43 +0000 (0:00:10.979)       0:00:36.663 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:47:53 +0000 (0:00:10.908)       0:00:47.572 ******** ",
                                "=============================================================================== ",
                                "redis : Restart redis container ---------------------------------------- 10.98s",
                                "redis : Restart redis-sentinel container ------------------------------- 10.91s",
                                "redis : Copying over redis config files --------------------------------- 3.89s",
                                "redis : Copying over default config.json files -------------------------- 3.17s",
                                "service-check-containers : redis | Check containers --------------------- 3.14s",
                                "Group hosts based on enabled services ----------------------------------- 2.80s",
                                "redis : Ensuring config directories exist ------------------------------- 2.25s",
                                "Group hosts based on Kolla action --------------------------------------- 2.09s",
                                "service-check-containers : Include tasks -------------------------------- 1.93s",
                                "redis : include_tasks --------------------------------------------------- 1.73s",
                                "redis : Flush handlers -------------------------------------------------- 1.72s",
                                "service-check-containers : redis | Notify handlers to restart containers --- 1.61s",
                                "+ osism apply -a upgrade mariadb",
                                "\u001b[32m2026-03-26 04:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44634b2e-5b51-4f9a-9bb1-cb1a991037c4 (mariadb) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 44634b2e-5b51-4f9a-9bb1-cb1a991037c4 (mariadb) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:48:02 +0000 (0:00:01.330)       0:00:01.330 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:48:04 +0000 (0:00:02.378)       0:00:03.708 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Thursday 26 March 2026  04:48:07 +0000 (0:00:02.678)       0:00:06.387 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  04:48:08 +0000 (0:00:01.508)       0:00:07.895 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Thursday 26 March 2026  04:48:10 +0000 (0:00:01.753)       0:00:09.649 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Thursday 26 March 2026  04:48:14 +0000 (0:00:03.606)       0:00:13.255 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Thursday 26 March 2026  04:48:15 +0000 (0:00:01.606)       0:00:14.862 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Thursday 26 March 2026  04:48:18 +0000 (0:00:02.251)       0:00:17.116 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Thursday 26 March 2026  04:48:22 +0000 (0:00:04.504)       0:00:21.621 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Thursday 26 March 2026  04:48:24 +0000 (0:00:02.049)       0:00:23.671 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  04:48:29 +0000 (0:00:04.773)       0:00:28.444 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Thursday 26 March 2026  04:48:31 +0000 (0:00:01.881)       0:00:30.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:48:34 +0000 (0:00:03.572)       0:00:33.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Thursday 26 March 2026  04:48:38 +0000 (0:00:03.520)       0:00:37.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : mariadb | Check containers] *******************",
                                "Thursday 26 March 2026  04:48:42 +0000 (0:00:04.149)       0:00:41.569 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : mariadb | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:48:46 +0000 (0:00:04.154)       0:00:45.723 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:48:48 +0000 (0:00:01.392)       0:00:47.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Checking for mariadb cluster] **********************************",
                                "Thursday 26 March 2026  04:48:52 +0000 (0:00:04.186)       0:00:51.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Cleaning up temp file on localhost] ****************************",
                                "Thursday 26 March 2026  04:48:53 +0000 (0:00:01.400)       0:00:52.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Stop MariaDB containers] ***************************************",
                                "Thursday 26 March 2026  04:48:54 +0000 (0:00:01.131)       0:00:53.835 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Run MariaDB wsrep recovery] ************************************",
                                "Thursday 26 March 2026  04:48:56 +0000 (0:00:01.391)       0:00:55.226 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying MariaDB log file to /tmp] ******************************",
                                "Thursday 26 March 2026  04:48:57 +0000 (0:00:01.536)       0:00:56.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB wsrep recovery seqno] ******************************",
                                "Thursday 26 March 2026  04:48:59 +0000 (0:00:01.358)       0:00:58.122 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Removing MariaDB log file from /tmp] ***************************",
                                "Thursday 26 March 2026  04:49:01 +0000 (0:00:01.927)       0:01:00.050 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Registering MariaDB seqno variable] ****************************",
                                "Thursday 26 March 2026  04:49:02 +0000 (0:00:01.357)       0:01:01.407 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Comparing seqno value on all mariadb hosts] ********************",
                                "Thursday 26 March 2026  04:49:03 +0000 (0:00:01.545)       0:01:02.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Writing hostname of host with the largest seqno to temp file] ***",
                                "Thursday 26 March 2026  04:49:05 +0000 (0:00:01.507)       0:01:04.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Registering mariadb_recover_inventory_name from temp file] *****",
                                "Thursday 26 March 2026  04:49:06 +0000 (0:00:01.368)       0:01:05.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap and master hostnames into facts] ***************",
                                "Thursday 26 March 2026  04:49:08 +0000 (0:00:01.405)       0:01:07.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Set grastate.dat file from MariaDB container in bootstrap host] ***",
                                "Thursday 26 March 2026  04:49:09 +0000 (0:00:01.336)       0:01:08.571 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Starting first MariaDB container] ******************************",
                                "Thursday 26 March 2026  04:49:10 +0000 (0:00:01.377)       0:01:09.949 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for first MariaDB container] ******************************",
                                "Thursday 26 March 2026  04:49:12 +0000 (0:00:01.332)       0:01:11.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Set first MariaDB container as primary] ************************",
                                "Thursday 26 March 2026  04:49:13 +0000 (0:00:01.568)       0:01:12.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB to become operational] ************************",
                                "Thursday 26 March 2026  04:49:15 +0000 (0:00:01.475)       0:01:14.325 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Restart slave MariaDB container(s)] ****************************",
                                "Thursday 26 March 2026  04:49:16 +0000 (0:00:01.400)       0:01:15.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for slave MariaDB] ****************************************",
                                "Thursday 26 March 2026  04:49:20 +0000 (0:00:03.569)       0:01:19.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Restart master MariaDB container(s)] ***************************",
                                "Thursday 26 March 2026  04:49:21 +0000 (0:00:01.582)       0:01:20.878 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/mariadb-server:10.11.15.20251208', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for master mariadb] ***************************************",
                                "Thursday 26 March 2026  04:49:25 +0000 (0:00:03.409)       0:01:24.287 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Get container facts] ***************************",
                                "Thursday 26 March 2026  04:49:26 +0000 (0:00:01.730)       0:01:26.018 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are missing or not running] ***",
                                "Thursday 26 March 2026  04:49:28 +0000 (0:00:01.362)       0:01:27.380 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are unhealthy] **************",
                                "Thursday 26 March 2026  04:49:29 +0000 (0:00:01.330)       0:01:28.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Thursday 26 March 2026  04:49:31 +0000 (0:00:01.801)       0:01:30.512 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Thursday 26 March 2026  04:49:33 +0000 (0:00:01.951)       0:01:32.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Thursday 26 March 2026  04:49:35 +0000 (0:00:01.959)       0:01:34.423 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Thursday 26 March 2026  04:49:36 +0000 (0:00:01.336)       0:01:35.760 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Thursday 26 March 2026  04:49:38 +0000 (0:00:01.387)       0:01:37.147 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Thursday 26 March 2026  04:49:40 +0000 (0:00:02.028)       0:01:39.176 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Thursday 26 March 2026  04:49:41 +0000 (0:00:01.443)       0:01:40.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Thursday 26 March 2026  04:49:42 +0000 (0:00:01.379)       0:01:41.999 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Thursday 26 March 2026  04:49:46 +0000 (0:00:03.524)       0:01:45.523 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Thursday 26 March 2026  04:49:47 +0000 (0:00:01.398)       0:01:46.921 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Thursday 26 March 2026  04:49:49 +0000 (0:00:01.455)       0:01:48.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  04:49:51 +0000 (0:00:01.793)       0:01:50.171 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Thursday 26 March 2026  04:49:52 +0000 (0:00:01.537)       0:01:51.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Thursday 26 March 2026  04:49:54 +0000 (0:00:01.619)       0:01:53.329 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Thursday 26 March 2026  04:49:55 +0000 (0:00:01.685)       0:01:55.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  04:49:58 +0000 (0:00:02.231)       0:01:57.246 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  04:50:24 +0000 (0:00:26.258)       0:02:23.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  04:50:29 +0000 (0:00:04.617)       0:02:28.123 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  04:50:32 +0000 (0:00:02.976)       0:02:31.099 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  04:50:58 +0000 (0:00:26.022)       0:02:57.122 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  04:51:02 +0000 (0:00:04.564)       0:03:01.686 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Thursday 26 March 2026  04:51:05 +0000 (0:00:02.870)       0:03:04.557 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Thursday 26 March 2026  04:51:30 +0000 (0:00:25.220)       0:03:29.777 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Wait for MariaDB service port liveness (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Thursday 26 March 2026  04:51:38 +0000 (0:00:07.967)       0:03:37.745 ******** ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Thursday 26 March 2026  04:51:42 +0000 (0:00:03.948)       0:03:41.693 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Thursday 26 March 2026  04:51:44 +0000 (0:00:01.988)       0:03:43.681 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Thursday 26 March 2026  04:51:47 +0000 (0:00:03.098)       0:03:46.780 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Thursday 26 March 2026  04:51:51 +0000 (0:00:03.287)       0:03:50.067 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Thursday 26 March 2026  04:51:54 +0000 (0:00:03.181)       0:03:53.248 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Get container facts] ***************************",
                                "Thursday 26 March 2026  04:51:57 +0000 (0:00:03.482)       0:03:56.731 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are missing or not running] ***",
                                "Thursday 26 March 2026  04:52:04 +0000 (0:00:06.649)       0:04:03.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are unhealthy] **************",
                                "Thursday 26 March 2026  04:52:08 +0000 (0:00:03.812)       0:04:07.193 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Thursday 26 March 2026  04:52:09 +0000 (0:00:01.604)       0:04:08.798 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Thursday 26 March 2026  04:52:13 +0000 (0:00:03.415)       0:04:12.214 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Run upgrade in MariaDB container] ******************************",
                                "Thursday 26 March 2026  04:52:15 +0000 (0:00:01.996)       0:04:14.210 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:52:32 +0000 (0:00:17.715)       0:04:31.925 ******** ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 77.50s",
                                "mariadb : Run upgrade in MariaDB container ----------------------------- 17.72s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 17.15s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 9.79s",
                                "service-check : mariadb | Get container facts --------------------------- 6.65s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 4.77s",
                                "mariadb : Copying over config.json files for services ------------------- 4.51s",
                                "service-check-containers : Include tasks -------------------------------- 4.19s",
                                "service-check-containers : mariadb | Check containers ------------------- 4.15s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 4.15s",
                                "service-check : mariadb | Fail if containers are missing or not running --- 3.81s",
                                "mariadb : Ensuring config directories exist ----------------------------- 3.61s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 3.57s",
                                "mariadb : Restart slave MariaDB container(s) ---------------------------- 3.57s",
                                "mariadb : Check MariaDB service WSREP sync status ----------------------- 3.52s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 3.52s",
                                "mariadb : Granting permissions on Mariabackup database to backup user --- 3.48s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 3.42s",
                                "mariadb : Restart master MariaDB container(s) --------------------------- 3.41s",
                                "mariadb : Creating mysql monitor user ----------------------------------- 3.29s",
                                "+ osism apply -a upgrade rabbitmq",
                                "\u001b[32m2026-03-26 04:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e3e4e19-e6c3-4dbe-b118-c2b62a5d823d (rabbitmq) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2e3e4e19-e6c3-4dbe-b118-c2b62a5d823d (rabbitmq) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:52:41 +0000 (0:00:01.444)       0:00:01.444 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:52:43 +0000 (0:00:01.871)       0:00:03.315 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  04:52:45 +0000 (0:00:01.751)       0:00:05.067 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Thursday 26 March 2026  04:52:47 +0000 (0:00:02.804)       0:00:07.871 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Thursday 26 March 2026  04:52:50 +0000 (0:00:02.382)       0:00:10.254 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Thursday 26 March 2026  04:52:53 +0000 (0:00:03.398)       0:00:13.653 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Thursday 26 March 2026  04:53:02 +0000 (0:00:09.086)       0:00:22.740 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Thursday 26 March 2026  04:53:04 +0000 (0:00:01.301)       0:00:24.041 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  04:53:05 +0000 (0:00:01.725)       0:00:25.767 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Thursday 26 March 2026  04:53:07 +0000 (0:00:01.906)       0:00:27.674 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Thursday 26 March 2026  04:53:09 +0000 (0:00:02.306)       0:00:29.981 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Thursday 26 March 2026  04:53:12 +0000 (0:00:02.750)       0:00:32.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Thursday 26 March 2026  04:53:14 +0000 (0:00:01.877)       0:00:34.609 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Thursday 26 March 2026  04:53:16 +0000 (0:00:01.771)       0:00:36.380 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Thursday 26 March 2026  04:53:18 +0000 (0:00:02.399)       0:00:38.780 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Thursday 26 March 2026  04:53:21 +0000 (0:00:02.454)       0:00:41.235 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Thursday 26 March 2026  04:53:24 +0000 (0:00:03.109)       0:00:44.344 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Thursday 26 March 2026  04:53:26 +0000 (0:00:02.426)       0:00:46.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Thursday 26 March 2026  04:53:29 +0000 (0:00:02.314)       0:00:49.085 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Thursday 26 March 2026  04:53:31 +0000 (0:00:02.290)       0:00:51.376 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Thursday 26 March 2026  04:53:33 +0000 (0:00:02.591)       0:00:53.968 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over extra CA certificates] *******",
                                "Thursday 26 March 2026  04:53:35 +0000 (0:00:01.793)       0:00:55.761 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over backend internal TLS certificate] ***",
                                "Thursday 26 March 2026  04:53:38 +0000 (0:00:02.341)       0:00:58.103 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over backend internal TLS key] ****",
                                "Thursday 26 March 2026  04:53:39 +0000 (0:00:01.452)       0:00:59.555 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Thursday 26 March 2026  04:53:41 +0000 (0:00:01.778)       0:01:01.334 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : rabbitmq | Check containers] ******************",
                                "Thursday 26 March 2026  04:53:45 +0000 (0:00:03.710)       0:01:05.044 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : rabbitmq | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:53:47 +0000 (0:00:02.187)       0:01:07.231 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:53:48 +0000 (0:00:01.416)       0:01:08.648 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/rabbitmq:4.1.5.20251208', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Thursday 26 March 2026  04:53:50 +0000 (0:00:02.091)       0:01:10.740 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  04:53:52 +0000 (0:00:01.778)       0:01:12.518 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  04:53:54 +0000 (0:00:02.163)       0:01:14.682 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  04:54:03 +0000 (0:00:09.146)       0:01:23.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  04:54:12 +0000 (0:00:09.086)       0:01:32.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  04:54:21 +0000 (0:00:08.807)       0:01:41.722 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  04:54:23 +0000 (0:00:01.683)       0:01:43.405 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  04:54:31 +0000 (0:00:08.028)       0:01:51.434 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  04:54:45 +0000 (0:00:13.625)       0:02:05.060 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Thursday 26 March 2026  04:54:53 +0000 (0:00:08.243)       0:02:13.303 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Thursday 26 March 2026  04:54:55 +0000 (0:00:01.729)       0:02:15.033 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Thursday 26 March 2026  04:55:03 +0000 (0:00:08.941)       0:02:23.974 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Thursday 26 March 2026  04:55:18 +0000 (0:00:14.227)       0:02:38.202 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Thursday 26 March 2026  04:55:28 +0000 (0:00:10.661)       0:02:48.864 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Thursday 26 March 2026  04:55:30 +0000 (0:00:01.517)       0:02:50.382 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:55:34 +0000 (0:00:04.463)       0:02:54.846 ******** ",
                                "=============================================================================== ",
                                "rabbitmq : Restart rabbitmq container ---------------------------------- 36.94s",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------- 27.71s",
                                "rabbitmq : Put RabbitMQ node into maintenance mode --------------------- 26.12s",
                                "rabbitmq : Get new RabbitMQ version ------------------------------------- 9.09s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 5.58s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 4.46s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 3.71s",
                                "rabbitmq : Get current RabbitMQ version --------------------------------- 3.40s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 3.11s",
                                "rabbitmq : include_tasks ------------------------------------------------ 2.80s",
                                "rabbitmq : List RabbitMQ policies --------------------------------------- 2.75s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 2.59s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 2.45s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 2.43s",
                                "rabbitmq : Copying over config.json files for services ------------------ 2.40s",
                                "rabbitmq : Get container facts ------------------------------------------ 2.38s",
                                "service-cert-copy : rabbitmq | Copying over extra CA certificates ------- 2.34s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 2.32s",
                                "rabbitmq : Get container facts ------------------------------------------ 2.31s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 2.29s",
                                "+ osism apply -a upgrade openvswitch",
                                "\u001b[32m2026-03-26 04:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3fbc17bd-68e1-4e2f-bde2-7bb54783be9e (openvswitch) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3fbc17bd-68e1-4e2f-bde2-7bb54783be9e (openvswitch) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x739f147b4c20>): Expecting value: line 2 column 1 (char 1)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x739f147b4c20>): 'NoneType' object is not subscriptable\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:55:42 +0000 (0:00:01.079)       0:00:01.079 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:55:44 +0000 (0:00:01.553)       0:00:02.633 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Thursday 26 March 2026  04:55:45 +0000 (0:00:01.072)       0:00:03.706 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Thursday 26 March 2026  04:55:47 +0000 (0:00:02.150)       0:00:05.856 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Thursday 26 March 2026  04:55:49 +0000 (0:00:01.451)       0:00:07.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Thursday 26 March 2026  04:55:50 +0000 (0:00:01.386)       0:00:08.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Thursday 26 March 2026  04:55:52 +0000 (0:00:01.959)       0:00:10.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Thursday 26 March 2026  04:55:53 +0000 (0:00:01.091)       0:00:11.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Thursday 26 March 2026  04:55:55 +0000 (0:00:01.717)       0:00:13.463 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Thursday 26 March 2026  04:55:57 +0000 (0:00:02.390)       0:00:15.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-check-containers : openvswitch | Check containers] ***************",
                                "Thursday 26 March 2026  04:55:59 +0000 (0:00:01.442)       0:00:17.296 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : openvswitch | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:56:01 +0000 (0:00:02.097)       0:00:19.393 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:56:02 +0000 (0:00:00.936)       0:00:20.330 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_handler_task_start) in callback\u001b[0m",
                                "\u001b[1;35mplugin (<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x739f147b4c20>): 'NoneType' object is not subscriptable\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-db-server:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/2025.1/openvswitch-vswitchd:3.5.1.20251208', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:03 +0000 (0:00:01.833)       0:00:22.164 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:04 +0000 (0:00:00.164)       0:00:22.329 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:04 +0000 (0:00:00.153)       0:00:22.483 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:04 +0000 (0:00:00.144)       0:00:22.627 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:04 +0000 (0:00:00.348)       0:00:22.975 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Thursday 26 March 2026  04:56:04 +0000 (0:00:00.145)       0:00:23.121 ******** ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Thursday 26 March 2026  04:56:05 +0000 (0:00:00.150)       0:00:23.271 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Thursday 26 March 2026  04:56:15 +0000 (0:00:10.909)       0:00:34.180 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Thursday 26 March 2026  04:56:17 +0000 (0:00:01.206)       0:00:35.387 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Thursday 26 March 2026  04:56:27 +0000 (0:00:10.486)       0:00:45.873 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Thursday 26 March 2026  04:56:34 +0000 (0:00:06.323)       0:00:52.197 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Thursday 26 March 2026  04:56:36 +0000 (0:00:02.279)       0:00:54.476 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  04:56:40 +0000 (0:00:04.022)       0:00:58.498 ******** ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-db-server container ------------------ 10.91s",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 10.49s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 6.32s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 4.02s",
                                "openvswitch : Copying over config.json files for services --------------- 2.39s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 2.28s",
                                "openvswitch : include_tasks --------------------------------------------- 2.15s",
                                "service-check-containers : openvswitch | Check containers --------------- 2.10s",
                                "module-load : Drop module persistence ----------------------------------- 1.96s",
                                "service-check-containers : Include tasks -------------------------------- 1.83s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.72s",
                                "Group hosts based on Kolla action --------------------------------------- 1.55s",
                                "module-load : Load modules ---------------------------------------------- 1.45s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 1.44s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.39s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.21s",
                                "openvswitch : Flush Handlers -------------------------------------------- 1.11s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 1.09s",
                                "Group hosts based on enabled services ----------------------------------- 1.07s",
                                "service-check-containers : openvswitch | Notify handlers to restart containers --- 0.94s",
                                "+ osism apply -a upgrade ovn",
                                "\u001b[32m2026-03-26 04:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 53508c48-b0a1-45e6-8522-6a5123067a74 (ovn) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 04:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 53508c48-b0a1-45e6-8522-6a5123067a74 (ovn) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Thursday 26 March 2026  04:56:48 +0000 (0:00:01.326)       0:00:01.326 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Thursday 26 March 2026  04:56:52 +0000 (0:00:03.418)       0:00:04.744 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Thursday 26 March 2026  04:56:55 +0000 (0:00:03.073)       0:00:07.818 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Thursday 26 March 2026  04:56:58 +0000 (0:00:02.771)       0:00:10.590 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Thursday 26 March 2026  04:57:00 +0000 (0:00:02.402)       0:00:12.993 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Thursday 26 March 2026  04:57:03 +0000 (0:00:02.655)       0:00:15.648 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Thursday 26 March 2026  04:57:05 +0000 (0:00:02.288)       0:00:17.937 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_controller | Check containers] ************",
                                "Thursday 26 March 2026  04:57:08 +0000 (0:00:03.144)       0:00:21.081 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_controller | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  04:57:11 +0000 (0:00:02.529)       0:00:23.611 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  04:57:13 +0000 (0:00:02.009)       0:00:25.621 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-controller:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Thursday 26 March 2026  04:57:15 +0000 (0:00:02.506)       0:00:28.127 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Thursday 26 March 2026  04:57:19 +0000 (0:00:03.550)       0:00:31.677 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:39 +0000 (0:00:20.184)       0:00:51.861 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:39 +0000 (0:00:00.458)       0:00:52.320 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:40 +0000 (0:00:00.464)       0:00:52.785 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:40 +0000 (0:00:00.445)       0:00:53.231 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:41 +0000 (0:00:00.450)       0:00:53.681 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Thursday 26 March 2026  04:57:41 +0000 (0:00:00.457)       0:00:54.139 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Thursday 26 March 2026  04:57:42 +0000 (0:00:00.841)       0:00:54.980 ******** ",
                                "",
                                "STILL ALIVE [task 'ovn-controller : Restart ovn-controller container' is running] ***",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:59:54 +0000 (0:02:11.773)       0:03:06.754 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/upgrade.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  04:59:56 +0000 (0:00:01.950)       0:03:08.705 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Thursday 26 March 2026  04:59:58 +0000 (0:00:01.933)       0:03:10.639 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Thursday 26 March 2026  04:59:59 +0000 (0:00:01.802)       0:03:12.442 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Thursday 26 March 2026  05:00:01 +0000 (0:00:01.355)       0:03:13.797 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Thursday 26 March 2026  05:00:02 +0000 (0:00:01.418)       0:03:15.216 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Thursday 26 March 2026  05:00:04 +0000 (0:00:01.650)       0:03:16.866 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Thursday 26 March 2026  05:00:05 +0000 (0:00:01.423)       0:03:18.290 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Thursday 26 March 2026  05:00:07 +0000 (0:00:01.482)       0:03:19.772 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Thursday 26 March 2026  05:00:08 +0000 (0:00:01.687)       0:03:21.459 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Thursday 26 March 2026  05:00:10 +0000 (0:00:01.505)       0:03:22.965 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Thursday 26 March 2026  05:00:12 +0000 (0:00:01.863)       0:03:24.828 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Thursday 26 March 2026  05:00:13 +0000 (0:00:01.432)       0:03:26.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Thursday 26 March 2026  05:00:15 +0000 (0:00:01.333)       0:03:27.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Thursday 26 March 2026  05:00:16 +0000 (0:00:01.425)       0:03:29.019 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Thursday 26 March 2026  05:00:18 +0000 (0:00:01.814)       0:03:30.834 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Thursday 26 March 2026  05:00:19 +0000 (0:00:01.345)       0:03:32.180 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Thursday 26 March 2026  05:00:21 +0000 (0:00:02.167)       0:03:34.348 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Thursday 26 March 2026  05:00:23 +0000 (0:00:01.441)       0:03:35.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Thursday 26 March 2026  05:00:24 +0000 (0:00:01.411)       0:03:37.201 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Thursday 26 March 2026  05:00:26 +0000 (0:00:01.722)       0:03:38.923 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Thursday 26 March 2026  05:00:30 +0000 (0:00:04.227)       0:03:43.151 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Ensure configuration for relays exists] *************************",
                                "Thursday 26 March 2026  05:00:37 +0000 (0:00:06.531)       0:03:49.683 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****",
                                "Thursday 26 March 2026  05:00:39 +0000 (0:00:01.944)       0:03:51.627 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for OVN relay services] **********",
                                "Thursday 26 March 2026  05:00:40 +0000 (0:00:01.864)       0:03:53.492 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Generate config files for OVN relay services] *******************",
                                "Thursday 26 March 2026  05:00:43 +0000 (0:00:02.670)       0:03:56.163 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers] ********************",
                                "Thursday 26 March 2026  05:00:46 +0000 (0:00:03.022)       0:03:59.186 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  05:00:52 +0000 (0:00:05.498)       0:04:04.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Thursday 26 March 2026  05:00:53 +0000 (0:00:01.527)       0:04:06.212 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-northd:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-nb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-server:25.3.1.20251208', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/release/2025.1/ovn-sb-db-relay:25.3.1.20251208', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers with iteration] *****",
                                "Thursday 26 March 2026  05:00:57 +0000 (0:00:03.304)       0:04:09.517 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Thursday 26 March 2026  05:00:59 +0000 (0:00:02.246)       0:04:11.764 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  05:01:00 +0000 (0:00:01.561)       0:04:13.325 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  05:01:01 +0000 (0:00:00.424)       0:04:13.750 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Thursday 26 March 2026  05:01:01 +0000 (0:00:00.453)       0:04:14.203 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Thursday 26 March 2026  05:01:02 +0000 (0:00:01.021)       0:04:15.225 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Thursday 26 March 2026  05:01:19 +0000 (0:00:16.856)       0:04:32.081 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************",
                                "Thursday 26 March 2026  05:01:36 +0000 (0:00:16.931)       0:04:49.013 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Thursday 26 March 2026  05:01:47 +0000 (0:00:11.413)       0:05:00.426 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Thursday 26 March 2026  05:02:05 +0000 (0:00:17.732)       0:05:18.159 ******** ",
                                "Pausing for 5 seconds",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Thursday 26 March 2026  05:02:11 +0000 (0:00:06.175)       0:05:24.334 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Thursday 26 March 2026  05:02:13 +0000 (0:00:01.883)       0:05:26.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Thursday 26 March 2026  05:02:15 +0000 (0:00:01.825)       0:05:28.043 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Thursday 26 March 2026  05:02:17 +0000 (0:00:01.865)       0:05:29.908 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Thursday 26 March 2026  05:02:19 +0000 (0:00:01.640)       0:05:31.549 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Thursday 26 March 2026  05:02:20 +0000 (0:00:01.938)       0:05:33.487 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************",
                                "Thursday 26 March 2026  05:02:22 +0000 (0:00:01.912)       0:05:35.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=49  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=48  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=47  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  05:02:26 +0000 (0:00:03.159)       0:05:38.559 ******** ",
                                "=============================================================================== ",
                                "ovn-controller : Restart ovn-controller container --------------------- 131.77s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 20.18s",
                                "ovn-db : Restart ovn-northd container ---------------------------------- 17.73s",
                                "ovn-db : Restart ovn-sb-db container ----------------------------------- 16.93s",
                                "ovn-db : Restart ovn-nb-db container ----------------------------------- 16.86s",
                                "ovn-db : Restart ovn-sb-db-relay container ----------------------------- 11.41s",
                                "ovn-db : Copying over config.json files for services -------------------- 6.53s",
                                "ovn-db : Wait for leader election --------------------------------------- 6.18s",
                                "service-check-containers : ovn_db | Check containers -------------------- 5.50s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 4.23s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 3.55s",
                                "Group hosts based on Kolla action --------------------------------------- 3.42s",
                                "service-check-containers : Include tasks -------------------------------- 3.30s",
                                "ovn-db : Wait for ovn-sb-db-relay --------------------------------------- 3.16s",
                                "ovn-controller : Copying over systemd override -------------------------- 3.14s",
                                "ovn-controller : Flush handlers ----------------------------------------- 3.12s",
                                "Group hosts based on enabled services ----------------------------------- 3.07s",
                                "ovn-db : Generate config files for OVN relay services ------------------- 3.02s",
                                "ovn-controller : include_tasks ------------------------------------------ 2.77s",
                                "ovn-db : Copying over config.json files for OVN relay services ---------- 2.67s",
                                "+ [[ false == \\f\\a\\l\\s\\e ]]",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ osism apply ceph-rolling_update -e ireallymeanit=yes",
                                "\u001b[32m2026-03-26 05:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3cc48515-9882-4eae-9b27-8ff8065ea8e3 (ceph-rolling_update) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 05:02:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3cc48515-9882-4eae-9b27-8ff8065ea8e3 (ceph-rolling_update) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Confirm whether user really meant to upgrade the cluster] ****************",
                                "",
                                "TASK [Exit playbook, if user did not mean to upgrade cluster] ******************",
                                "Thursday 26 March 2026  05:02:38 +0000 (0:00:01.792)       0:00:01.792 ******** ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: rbdmirrors\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: nfss\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: clients\u001b[0m",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "PLAY [Gather facts and check the init system] **********************************",
                                "",
                                "TASK [Gather facts on all Ceph hosts for following reference] ******************",
                                "Thursday 26 March 2026  05:02:39 +0000 (0:00:01.712)       0:00:03.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"gather facts on all Ceph hosts for following reference\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather facts] ************************************************************",
                                "Thursday 26 March 2026  05:02:47 +0000 (0:00:07.308)       0:00:10.813 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Gather and delegate facts] ***********************************************",
                                "Thursday 26 March 2026  05:02:53 +0000 (0:00:06.298)       0:00:17.111 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [Set_fact rolling_update] *************************************************",
                                "Thursday 26 March 2026  05:03:24 +0000 (0:00:31.499)       0:00:48.611 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:03:27 +0000 (0:00:02.108)       0:00:50.719 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:03:29 +0000 (0:00:02.740)       0:00:53.460 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:03:32 +0000 (0:00:02.663)       0:00:56.123 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:03:34 +0000 (0:00:01.965)       0:00:58.089 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:03:36 +0000 (0:00:02.538)       0:01:00.627 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:03:38 +0000 (0:00:01.987)       0:01:02.615 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:03:41 +0000 (0:00:02.137)       0:01:04.752 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:03:43 +0000 (0:00:01.905)       0:01:06.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:03:45 +0000 (0:00:02.534)       0:01:09.192 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:03:47 +0000 (0:00:02.237)       0:01:11.430 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:03:49 +0000 (0:00:01.661)       0:01:13.092 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:03:51 +0000 (0:00:02.119)       0:01:15.211 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:03:54 +0000 (0:00:03.307)       0:01:18.519 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:03:56 +0000 (0:00:01.437)       0:01:19.956 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:03:58 +0000 (0:00:01.935)       0:01:21.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:03:59 +0000 (0:00:01.194)       0:01:23.086 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:03:52.185577', 'end': '2026-03-26 05:03:52.243863', 'delta': '0:00:00.058286', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:03:53.038758', 'end': '2026-03-26 05:03:53.088240', 'delta': '0:00:00.049482', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:03:53.633676', 'end': '2026-03-26 05:03:53.688108', 'delta': '0:00:00.054432', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:04:00 +0000 (0:00:01.242)       0:01:24.329 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:04:02 +0000 (0:00:02.174)       0:01:26.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:04:04 +0000 (0:00:01.279)       0:01:27.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:04:06 +0000 (0:00:02.152)       0:01:29.937 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:04:09 +0000 (0:00:03.367)       0:01:33.305 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:04:11 +0000 (0:00:02.171)       0:01:35.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:04:12 +0000 (0:00:01.175)       0:01:36.652 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:04:14 +0000 (0:00:01.207)       0:01:37.859 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:04:16 +0000 (0:00:02.518)       0:01:40.378 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:04:18 +0000 (0:00:01.888)       0:01:42.266 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:04:20 +0000 (0:00:02.085)       0:01:44.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:04:22 +0000 (0:00:01.949)       0:01:46.301 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:04:24 +0000 (0:00:02.102)       0:01:48.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:04:26 +0000 (0:00:02.017)       0:01:50.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:04:28 +0000 (0:00:02.134)       0:01:52.556 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-40-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1060', 'sectorsize': '2048', 'size': '530.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2ece1d7d', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:04:31 +0000 (0:00:02.526)       0:01:55.083 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-40-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1060', 'sectorsize': '2048', 'size': '530.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2ece1d7d', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1', 'scsi-SQEMU_QEMU_HARDDISK_2ece1d7d-b762-44e0-80cf-0ec8d4e65a06-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:04:33 +0000 (0:00:02.390)       0:01:57.474 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:04:36 +0000 (0:00:02.554)       0:02:00.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:04:38 +0000 (0:00:02.027)       0:02:02.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:04:40 +0000 (0:00:02.483)       0:02:04.540 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:04:42 +0000 (0:00:01.950)       0:02:06.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:04:45 +0000 (0:00:02.589)       0:02:09.080 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:04:47 +0000 (0:00:01.980)       0:02:11.061 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:04:51 +0000 (0:00:04.227)       0:02:15.289 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:04:53 +0000 (0:00:02.297)       0:02:17.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:04:55 +0000 (0:00:02.018)       0:02:19.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:04:57 +0000 (0:00:01.688)       0:02:21.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:04:59 +0000 (0:00:01.410)       0:02:22.704 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:05:00 +0000 (0:00:01.340)       0:02:24.045 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:05:01 +0000 (0:00:01.414)       0:02:25.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:05:03 +0000 (0:00:01.677)       0:02:27.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:05:05 +0000 (0:00:01.718)       0:02:28.856 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:05:06 +0000 (0:00:01.586)       0:02:30.443 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:05:08 +0000 (0:00:01.430)       0:02:31.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:05:09 +0000 (0:00:01.498)       0:02:33.372 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:05:11 +0000 (0:00:01.970)       0:02:35.343 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-infra : Update cache for Debian based OSs] **************************",
                                "Thursday 26 March 2026  05:05:14 +0000 (0:00:03.036)       0:02:38.380 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-infra : Include_tasks configure_firewall.yml] ***********************",
                                "Thursday 26 March 2026  05:05:22 +0000 (0:00:08.251)       0:02:46.631 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-infra : Include_tasks setup_ntp.yml] ********************************",
                                "Thursday 26 March 2026  05:05:25 +0000 (0:00:02.179)       0:02:48.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-infra : Add logrotate configuration] ********************************",
                                "Thursday 26 March 2026  05:05:27 +0000 (0:00:01.911)       0:02:50.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_system.yml] ********************************",
                                "Thursday 26 March 2026  05:05:30 +0000 (0:00:03.105)       0:02:53.828 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_system.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported ansible version (1.X)] ***************",
                                "Thursday 26 March 2026  05:05:33 +0000 (0:00:02.913)       0:02:56.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported system] ******************************",
                                "Thursday 26 March 2026  05:05:34 +0000 (0:00:01.808)       0:02:58.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported architecture] ************************",
                                "Thursday 26 March 2026  05:05:37 +0000 (0:00:02.111)       0:03:00.662 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported distribution] ************************",
                                "Thursday 26 March 2026  05:05:39 +0000 (0:00:02.036)       0:03:02.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported CentOS release] **********************",
                                "Thursday 26 March 2026  05:05:41 +0000 (0:00:02.169)       0:03:04.867 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported distribution for ubuntu cloud archive] ***",
                                "Thursday 26 March 2026  05:05:43 +0000 (0:00:01.955)       0:03:06.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail on unsupported SUSE/openSUSE distribution (only 15.x supported)] ***",
                                "Thursday 26 March 2026  05:05:45 +0000 (0:00:02.173)       0:03:08.997 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if systemd is not present] **************************",
                                "Thursday 26 March 2026  05:05:47 +0000 (0:00:01.947)       0:03:10.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate repository variables in non-containerized scenario] ***",
                                "Thursday 26 March 2026  05:05:49 +0000 (0:00:02.267)       0:03:13.212 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate osd_objectstore] ********************************",
                                "Thursday 26 March 2026  05:05:51 +0000 (0:00:02.082)       0:03:15.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate radosgw network configuration] ******************",
                                "Thursday 26 March 2026  05:05:53 +0000 (0:00:02.175)       0:03:17.470 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate lvm osd scenario] *******************************",
                                "Thursday 26 March 2026  05:05:56 +0000 (0:00:02.326)       0:03:19.796 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate bluestore lvm osd scenario] *********************",
                                "Thursday 26 March 2026  05:05:58 +0000 (0:00:01.940)       0:03:21.737 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if local scenario is enabled on debian] *************",
                                "Thursday 26 March 2026  05:06:00 +0000 (0:00:02.288)       0:03:24.025 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if rhcs repository is enabled on debian] ************",
                                "Thursday 26 March 2026  05:06:02 +0000 (0:00:01.980)       0:03:26.005 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check ceph_origin definition on SUSE/openSUSE Leap] ******",
                                "Thursday 26 March 2026  05:06:04 +0000 (0:00:02.178)       0:03:28.184 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check ceph_repository definition on SUSE/openSUSE Leap] ***",
                                "Thursday 26 March 2026  05:06:06 +0000 (0:00:01.996)       0:03:30.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate ntp daemon type] ********************************",
                                "Thursday 26 March 2026  05:06:08 +0000 (0:00:02.413)       0:03:32.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Abort if ntp_daemon_type is ntpd on Atomic] **************",
                                "Thursday 26 March 2026  05:06:10 +0000 (0:00:02.030)       0:03:34.625 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_devices.yml] *******************************",
                                "Thursday 26 March 2026  05:06:13 +0000 (0:00:02.077)       0:03:36.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_devices.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-validate : Set_fact root_device] ************************************",
                                "Thursday 26 March 2026  05:06:15 +0000 (0:00:02.515)       0:03:39.218 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Resolve devices in lvm_volumes] **************************",
                                "Thursday 26 March 2026  05:06:17 +0000 (0:00:01.463)       0:03:40.681 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Set_fact lvm_volumes_data_devices] ***********************",
                                "Thursday 26 March 2026  05:06:18 +0000 (0:00:01.382)       0:03:42.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.data_vg is undefined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if root_device is passed in lvm_volumes or devices] ***",
                                "Thursday 26 March 2026  05:06:20 +0000 (0:00:01.660)       0:03:43.725 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Get devices information] *********************************",
                                "Thursday 26 March 2026  05:06:21 +0000 (0:00:01.449)       0:03:45.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if one of the devices is not a device] **************",
                                "Thursday 26 March 2026  05:06:22 +0000 (0:00:01.356)       0:03:46.531 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail when gpt header found on osd devices] ***************",
                                "Thursday 26 March 2026  05:06:24 +0000 (0:00:01.413)       0:03:47.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check data logical volume] *******************************",
                                "Thursday 26 March 2026  05:06:25 +0000 (0:00:01.325)       0:03:49.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'})\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if one of the data logical volume is not a device or doesn't exist] ***",
                                "Thursday 26 March 2026  05:06:28 +0000 (0:00:02.964)       0:03:52.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a/osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 955, 'dev': 6, 'nlink': 1, 'atime': 1774493776.0892901, 'mtime': 1774493776.08229, 'ctime': 1774493776.08229, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a/osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-e2623153-bc41-510f-8884-ef957bb96082/osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 965, 'dev': 6, 'nlink': 1, 'atime': 1774493795.0405843, 'mtime': 1774493795.0355842, 'ctime': 1774493795.0355842, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-e2623153-bc41-510f-8884-ef957bb96082/osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-a652979e-9f40-503a-bbc8-6de5e605991e/osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 957, 'dev': 6, 'nlink': 1, 'atime': 1774493778.800435, 'mtime': 1774493778.795435, 'ctime': 1774493778.795435, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-a652979e-9f40-503a-bbc8-6de5e605991e/osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-b5eee7c3-8883-5bbe-be5a-75726e822543/osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 967, 'dev': 6, 'nlink': 1, 'atime': 1774493797.1287172, 'mtime': 1774493797.1227171, 'ctime': 1774493797.1227171, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-b5eee7c3-8883-5bbe-be5a-75726e822543/osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66/osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 949, 'dev': 6, 'nlink': 1, 'atime': 1774493778.807298, 'mtime': 1774493778.8022978, 'ctime': 1774493778.8022978, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64512, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66/osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'stat': {'exists': True, 'path': '/dev/ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771/osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'mode': '0660', 'isdir': False, 'ischr': False, 'isblk': True, 'isreg': False, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 6, 'size': 0, 'inode': 959, 'dev': 6, 'nlink': 1, 'atime': 1774493796.95358, 'mtime': 1774493796.9495797, 'ctime': 1774493796.9495797, 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': True, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': False, 'xoth': False, 'isuid': False, 'isgid': False, 'blocks': 0, 'block_size': 512, 'device_type': 64513, 'readable': True, 'writeable': True, 'executable': False, 'pw_name': 'root', 'gr_name': 'disk', 'mimetype': 'inode/symlink', 'charset': 'binary', 'version': None, 'attributes': [], 'attr_flags': ''}, 'invocation': {'module_args': {'path': '/dev/ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771/osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'follow': True, 'get_checksum': True, 'get_mime': True, 'get_attributes': True, 'checksum_algorithm': 'sha1'}}, 'failed': False, 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check bluestore db logical volume] ***********************",
                                "Thursday 26 March 2026  05:06:30 +0000 (0:00:01.595)       0:03:53.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if one of the bluestore db logical volume is not a device or doesn't exist] ***",
                                "Thursday 26 March 2026  05:06:31 +0000 (0:00:01.451)       0:03:55.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.db is defined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check bluestore wal logical volume] **********************",
                                "Thursday 26 March 2026  05:06:33 +0000 (0:00:01.497)       0:03:56.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if one of the bluestore wal logical volume is not a device or doesn't exist] ***",
                                "Thursday 26 March 2026  05:06:34 +0000 (0:00:01.663)       0:03:58.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a', 'data_vg': 'ceph-93e8c9a2-b6ff-5fe0-a79e-2922336c3e0a'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-e2623153-bc41-510f-8884-ef957bb96082', 'data_vg': 'ceph-e2623153-bc41-510f-8884-ef957bb96082'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-a652979e-9f40-503a-bbc8-6de5e605991e', 'data_vg': 'ceph-a652979e-9f40-503a-bbc8-6de5e605991e'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-b5eee7c3-8883-5bbe-be5a-75726e822543', 'data_vg': 'ceph-b5eee7c3-8883-5bbe-be5a-75726e822543'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-83c4def8-4703-5f7c-9549-7666ff9f2b66', 'data_vg': 'ceph-83c4def8-4703-5f7c-9549-7666ff9f2b66'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'item.wal is defined', 'item': {'data': 'osd-block-1fd8de68-da37-5e01-9bf2-5a04fcdcd771', 'data_vg': 'ceph-1fd8de68-da37-5e01-9bf2-5a04fcdcd771'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_eth_rgw.yml] *******************************",
                                "Thursday 26 March 2026  05:06:36 +0000 (0:00:01.384)       0:03:59.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_rgw_pools.yml] *****************************",
                                "Thursday 26 March 2026  05:06:38 +0000 (0:00:01.971)       0:04:01.800 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-validate/tasks/check_rgw_pools.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if ec_profile is not set for ec pools] **************",
                                "Thursday 26 March 2026  05:06:40 +0000 (0:00:02.635)       0:04:04.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if ec_k is not set for ec pools] ********************",
                                "Thursday 26 March 2026  05:06:42 +0000 (0:00:01.442)       0:04:05.877 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if ec_m is not set for ec pools] ********************",
                                "Thursday 26 March 2026  05:06:43 +0000 (0:00:01.713)       0:04:07.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_nfs.yml] ***********************************",
                                "Thursday 26 March 2026  05:06:45 +0000 (0:00:01.509)       0:04:09.100 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Include check_rbdmirror.yml] *****************************",
                                "Thursday 26 March 2026  05:06:47 +0000 (0:00:01.861)       0:04:10.961 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail if monitoring group doesn't exist] ******************",
                                "Thursday 26 March 2026  05:06:49 +0000 (0:00:02.202)       0:04:13.164 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail when monitoring doesn't contain at least one node.] ***",
                                "Thursday 26 March 2026  05:06:51 +0000 (0:00:02.090)       0:04:15.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Fail when dashboard_admin_password and/or grafana_admin_password are not set] ***",
                                "Thursday 26 March 2026  05:06:53 +0000 (0:00:02.016)       0:04:17.271 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate container registry credentials] *****************",
                                "Thursday 26 March 2026  05:06:55 +0000 (0:00:02.121)       0:04:19.393 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate container service and container package] ********",
                                "Thursday 26 March 2026  05:06:57 +0000 (0:00:02.059)       0:04:21.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate openstack_keys key format] **********************",
                                "Thursday 26 March 2026  05:07:00 +0000 (0:00:02.271)       0:04:23.724 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate clients keys key format] ************************",
                                "Thursday 26 March 2026  05:07:02 +0000 (0:00:02.175)       0:04:25.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate openstack_keys caps] ****************************",
                                "Thursday 26 March 2026  05:07:04 +0000 (0:00:02.159)       0:04:28.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.cinder-backup'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=vms, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.cinder'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=volumes, profile rbd pool=images'}, 'mode': '0600', 'name': 'client.glance'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=metrics'}, 'mode': '0600', 'name': 'client.gnocchi'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mon': 'profile rbd', 'osd': 'profile rbd pool=images, profile rbd pool=vms, profile rbd pool=volumes, profile rbd pool=backups'}, 'mode': '0600', 'name': 'client.nova'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'caps': {'mgr': 'allow rw', 'mon': 'allow r', 'osd': 'allow rw pool=cephfs_data'}, 'mode': '0600', 'name': 'client.manila'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate clients keys caps] ******************************",
                                "Thursday 26 March 2026  05:07:06 +0000 (0:00:02.193)       0:04:30.252 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Check virtual_ips is defined] ****************************",
                                "Thursday 26 March 2026  05:07:08 +0000 (0:00:02.309)       0:04:32.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-validate : Validate virtual_ips length] *****************************",
                                "Thursday 26 March 2026  05:07:10 +0000 (0:00:02.062)       0:04:34.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Include pre_requisites/prerequisites.yml] ********",
                                "Thursday 26 March 2026  05:07:13 +0000 (0:00:02.258)       0:04:36.883 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Include specific variables] **********************",
                                "Thursday 26 March 2026  05:07:16 +0000 (0:00:02.867)       0:04:39.750 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/ansible/roles/ceph-container-engine/vars/Debian.yml)\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Create the systemd docker override directory] ****",
                                "Thursday 26 March 2026  05:07:18 +0000 (0:00:02.053)       0:04:41.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Create the systemd docker override file] *********",
                                "Thursday 26 March 2026  05:07:20 +0000 (0:00:02.167)       0:04:43.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Remove docker proxy configuration] ***************",
                                "Thursday 26 March 2026  05:07:22 +0000 (0:00:01.993)       0:04:45.965 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Restart docker] **********************************",
                                "Thursday 26 March 2026  05:07:24 +0000 (0:00:02.604)       0:04:48.570 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Container registry authentication] ***************",
                                "Thursday 26 March 2026  05:07:27 +0000 (0:00:02.420)       0:04:50.990 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get the ceph release being deployed] *************************************",
                                "Thursday 26 March 2026  05:07:29 +0000 (0:00:02.355)       0:04:53.345 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Check ceph release being deployed] ***************************************",
                                "Thursday 26 March 2026  05:07:32 +0000 (0:00:02.637)       0:04:55.982 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Ensure cluster config is applied] ****************************************",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:07:33 +0000 (0:00:01.466)       0:04:57.450 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:07:35 +0000 (0:00:01.455)       0:04:58.905 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set cluster configs] *****************************************************",
                                "Thursday 26 March 2026  05:07:36 +0000 (0:00:01.150)       0:05:00.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mon cluster] ************************************************",
                                "",
                                "TASK [Remove ceph aliases] *****************************************************",
                                "Thursday 26 March 2026  05:07:46 +0000 (0:00:10.316)       0:05:10.373 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set mon_host_count] ******************************************************",
                                "Thursday 26 March 2026  05:07:48 +0000 (0:00:01.489)       0:05:11.862 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail when less than three monitors] **************************************",
                                "Thursday 26 March 2026  05:07:49 +0000 (0:00:01.151)       0:05:13.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Select a running monitor] ************************************************",
                                "Thursday 26 March 2026  05:07:50 +0000 (0:00:01.210)       0:05:14.224 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:07:51 +0000 (0:00:01.156)       0:05:15.381 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:07:52 +0000 (0:00:01.184)       0:05:16.566 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:07:54 +0000 (0:00:01.428)       0:05:17.994 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:07:55 +0000 (0:00:01.122)       0:05:19.117 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:07:56 +0000 (0:00:01.409)       0:05:20.527 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:07:58 +0000 (0:00:01.204)       0:05:21.732 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:07:59 +0000 (0:00:01.173)       0:05:22.905 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:08:00 +0000 (0:00:01.276)       0:05:24.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:08:01 +0000 (0:00:01.114)       0:05:25.296 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:08:02 +0000 (0:00:01.193)       0:05:26.489 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:08:04 +0000 (0:00:01.716)       0:05:28.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:08:05 +0000 (0:00:01.228)       0:05:29.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:08:09 +0000 (0:00:03.287)       0:05:32.722 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:08:10 +0000 (0:00:01.380)       0:05:34.102 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:08:12 +0000 (0:00:01.865)       0:05:35.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:08:13 +0000 (0:00:01.175)       0:05:37.144 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'c1b85917b265', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:08:06.328851', 'end': '2026-03-26 05:08:06.373241', 'delta': '0:00:00.044390', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['c1b85917b265'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:08:06.897478', 'end': '2026-03-26 05:08:06.954848', 'delta': '0:00:00.057370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:08:07.774569', 'end': '2026-03-26 05:08:07.828855', 'delta': '0:00:00.054286', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:08:14 +0000 (0:00:01.215)       0:05:38.360 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:08:16 +0000 (0:00:01.615)       0:05:39.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:08:17 +0000 (0:00:01.231)       0:05:41.207 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:08:18 +0000 (0:00:01.138)       0:05:42.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:08:20 +0000 (0:00:02.054)       0:05:44.401 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:08:21 +0000 (0:00:01.198)       0:05:45.599 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:08:23 +0000 (0:00:01.203)       0:05:46.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:08:24 +0000 (0:00:01.285)       0:05:48.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:08:25 +0000 (0:00:01.162)       0:05:49.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:08:26 +0000 (0:00:01.126)       0:05:50.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:08:27 +0000 (0:00:01.134)       0:05:51.512 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:08:28 +0000 (0:00:01.116)       0:05:52.628 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:08:30 +0000 (0:00:01.133)       0:05:53.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:08:31 +0000 (0:00:01.128)       0:05:54.891 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:08:32 +0000 (0:00:01.139)       0:05:56.031 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:08:33 +0000 (0:00:01.223)       0:05:57.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:08:34 +0000 (0:00:01.245)       0:05:58.500 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:08:36 +0000 (0:00:01.553)       0:06:00.054 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:08:37 +0000 (0:00:01.144)       0:06:01.199 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:08:39 +0000 (0:00:01.521)       0:06:02.720 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:08:40 +0000 (0:00:01.103)       0:06:03.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:08:41 +0000 (0:00:01.244)       0:06:05.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:08:42 +0000 (0:00:01.198)       0:06:06.267 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:08:44 +0000 (0:00:01.994)       0:06:08.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:08:45 +0000 (0:00:01.173)       0:06:09.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:08:46 +0000 (0:00:01.145)       0:06:10.581 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:08:49 +0000 (0:00:02.201)       0:06:12.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get ceph cluster status] *************************************************",
                                "Thursday 26 March 2026  05:08:52 +0000 (0:00:02.906)       0:06:15.689 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [Display ceph health detail] **********************************************",
                                "Thursday 26 March 2026  05:08:54 +0000 (0:00:02.285)       0:06:17.974 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail if cluster isn't in an acceptable state] ****************************",
                                "Thursday 26 March 2026  05:08:55 +0000 (0:00:01.227)       0:06:19.202 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get the ceph quorum status] **********************************************",
                                "Thursday 26 March 2026  05:08:56 +0000 (0:00:01.140)       0:06:20.342 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [Fail if the cluster quorum isn't in an acceptable state] *****************",
                                "Thursday 26 March 2026  05:08:58 +0000 (0:00:02.264)       0:06:22.607 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************",
                                "Thursday 26 March 2026  05:09:00 +0000 (0:00:01.162)       0:06:23.769 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [Create potentially missing keys (rbd and rbd-mirror)] ********************",
                                "Thursday 26 March 2026  05:09:02 +0000 (0:00:02.576)       0:06:26.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=['bootstrap-rbd', 'testbed-node-0'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=['bootstrap-rbd', 'testbed-node-1'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=['bootstrap-rbd', 'testbed-node-2'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=['bootstrap-rbd-mirror', 'testbed-node-0'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=['bootstrap-rbd-mirror', 'testbed-node-1'])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=['bootstrap-rbd-mirror', 'testbed-node-2'])\u001b[0m",
                                "",
                                "TASK [Stop ceph mon] ***********************************************************",
                                "Thursday 26 March 2026  05:09:16 +0000 (0:00:13.506)       0:06:39.853 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [Mask the mgr service] ****************************************************",
                                "Thursday 26 March 2026  05:09:19 +0000 (0:00:03.753)       0:06:43.607 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:09:22 +0000 (0:00:02.565)       0:06:46.173 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:09:24 +0000 (0:00:01.549)       0:06:47.723 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:09:25 +0000 (0:00:01.598)       0:06:49.322 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:09:27 +0000 (0:00:01.528)       0:06:50.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:09:28 +0000 (0:00:01.109)       0:06:51.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:09:29 +0000 (0:00:01.111)       0:06:53.071 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:09:30 +0000 (0:00:01.150)       0:06:54.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:09:32 +0000 (0:00:01.605)       0:06:55.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:09:33 +0000 (0:00:01.146)       0:06:56.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:09:34 +0000 (0:00:01.103)       0:06:58.078 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:09:36 +0000 (0:00:01.650)       0:06:59.729 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:09:37 +0000 (0:00:01.553)       0:07:01.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:09:38 +0000 (0:00:01.188)       0:07:02.471 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:09:39 +0000 (0:00:01.181)       0:07:03.653 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:09:41 +0000 (0:00:01.153)       0:07:04.807 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:09:42 +0000 (0:00:01.111)       0:07:05.918 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:09:43 +0000 (0:00:01.152)       0:07:07.070 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:09:44 +0000 (0:00:01.133)       0:07:08.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:09:45 +0000 (0:00:01.162)       0:07:09.366 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:09:46 +0000 (0:00:01.121)       0:07:10.488 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:09:47 +0000 (0:00:01.163)       0:07:11.652 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:09:49 +0000 (0:00:01.175)       0:07:12.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:09:50 +0000 (0:00:01.104)       0:07:13.933 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:09:51 +0000 (0:00:01.120)       0:07:15.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:09:52 +0000 (0:00:01.141)       0:07:16.195 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:09:53 +0000 (0:00:01.173)       0:07:17.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:09:54 +0000 (0:00:01.128)       0:07:18.497 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:09:56 +0000 (0:00:01.174)       0:07:19.671 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:09:57 +0000 (0:00:01.108)       0:07:20.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:09:58 +0000 (0:00:01.130)       0:07:21.909 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:09:59 +0000 (0:00:01.129)       0:07:23.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:10:00 +0000 (0:00:01.211)       0:07:24.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:10:01 +0000 (0:00:01.136)       0:07:25.386 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:10:02 +0000 (0:00:01.153)       0:07:26.540 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:10:04 +0000 (0:00:02.009)       0:07:28.549 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:10:07 +0000 (0:00:02.412)       0:07:30.962 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:10:08 +0000 (0:00:01.491)       0:07:32.454 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:10:09 +0000 (0:00:01.135)       0:07:33.590 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:10:11 +0000 (0:00:01.150)       0:07:34.740 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:10:13 +0000 (0:00:02.012)       0:07:36.752 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:10:14 +0000 (0:00:01.644)       0:07:38.396 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:10:15 +0000 (0:00:01.178)       0:07:39.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:10:17 +0000 (0:00:01.122)       0:07:40.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:10:18 +0000 (0:00:01.105)       0:07:41.803 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:10:19 +0000 (0:00:01.442)       0:07:43.245 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:10:21 +0000 (0:00:01.919)       0:07:45.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:10:22 +0000 (0:00:01.207)       0:07:46.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:10:23 +0000 (0:00:01.133)       0:07:47.506 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:10:25 +0000 (0:00:01.168)       0:07:48.674 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:10:26 +0000 (0:00:01.181)       0:07:49.856 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:10:27 +0000 (0:00:01.145)       0:07:51.001 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:10:28 +0000 (0:00:01.146)       0:07:52.148 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:10:31 +0000 (0:00:02.586)       0:07:54.734 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:10:32 +0000 (0:00:01.153)       0:07:55.888 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:10:33 +0000 (0:00:01.486)       0:07:57.375 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:10:34 +0000 (0:00:01.192)       0:07:58.568 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:10:36 +0000 (0:00:01.141)       0:07:59.710 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:10:37 +0000 (0:00:01.119)       0:08:00.829 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:10:38 +0000 (0:00:01.121)       0:08:01.951 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:10:39 +0000 (0:00:01.190)       0:08:03.142 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:10:40 +0000 (0:00:01.162)       0:08:04.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:10:41 +0000 (0:00:01.117)       0:08:05.422 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:10:42 +0000 (0:00:01.169)       0:08:06.591 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:10:44 +0000 (0:00:01.150)       0:08:07.741 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:10:45 +0000 (0:00:01.555)       0:08:09.297 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:10:52 +0000 (0:00:06.764)       0:08:16.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:10:53 +0000 (0:00:01.113)       0:08:17.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:10:54 +0000 (0:00:01.125)       0:08:18.301 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:10:55 +0000 (0:00:01.132)       0:08:19.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:10:56 +0000 (0:00:01.146)       0:08:20.581 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:10:58 +0000 (0:00:01.110)       0:08:21.691 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:10:59 +0000 (0:00:01.194)       0:08:22.885 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:11:00 +0000 (0:00:01.126)       0:08:24.012 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:11:01 +0000 (0:00:01.116)       0:08:25.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:11:02 +0000 (0:00:01.154)       0:08:26.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:11:03 +0000 (0:00:01.115)       0:08:27.399 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:11:04 +0000 (0:00:01.158)       0:08:28.558 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:11:06 +0000 (0:00:01.164)       0:08:29.722 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:11:07 +0000 (0:00:01.257)       0:08:30.980 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:11:08 +0000 (0:00:01.165)       0:08:32.147 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:11:09 +0000 (0:00:01.213)       0:08:33.360 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:11:10 +0000 (0:00:01.193)       0:08:34.554 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:11:12 +0000 (0:00:01.190)       0:08:35.744 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:11:13 +0000 (0:00:01.119)       0:08:36.864 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:11:14 +0000 (0:00:01.194)       0:08:38.059 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:11:15 +0000 (0:00:01.131)       0:08:39.190 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:11:16 +0000 (0:00:01.137)       0:08:40.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:11:18 +0000 (0:00:01.768)       0:08:42.096 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:11:19 +0000 (0:00:01.440)       0:08:43.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:11:21 +0000 (0:00:01.477)       0:08:45.015 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:11:22 +0000 (0:00:01.150)       0:08:46.166 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:11:23 +0000 (0:00:01.405)       0:08:47.571 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:11:25 +0000 (0:00:01.819)       0:08:49.391 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Thursday 26 March 2026  05:11:26 +0000 (0:00:01.148)       0:08:50.540 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Thursday 26 March 2026  05:11:28 +0000 (0:00:01.527)       0:08:52.068 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Thursday 26 March 2026  05:11:31 +0000 (0:00:03.507)       0:08:55.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Thursday 26 March 2026  05:11:33 +0000 (0:00:01.206)       0:08:56.781 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Thursday 26 March 2026  05:11:34 +0000 (0:00:01.146)       0:08:57.928 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Thursday 26 March 2026  05:11:35 +0000 (0:00:01.166)       0:08:59.095 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Thursday 26 March 2026  05:11:37 +0000 (0:00:02.085)       0:09:01.180 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Thursday 26 March 2026  05:11:39 +0000 (0:00:01.613)       0:09:02.794 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Thursday 26 March 2026  05:11:40 +0000 (0:00:01.465)       0:09:04.260 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Thursday 26 March 2026  05:11:42 +0000 (0:00:01.542)       0:09:05.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Thursday 26 March 2026  05:11:43 +0000 (0:00:01.688)       0:09:07.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Thursday 26 March 2026  05:11:45 +0000 (0:00:01.696)       0:09:09.188 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Thursday 26 March 2026  05:11:49 +0000 (0:00:03.888)       0:09:13.076 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Thursday 26 March 2026  05:11:51 +0000 (0:00:02.132)       0:09:15.209 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Thursday 26 March 2026  05:11:52 +0000 (0:00:01.155)       0:09:16.365 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Thursday 26 March 2026  05:11:53 +0000 (0:00:01.179)       0:09:17.544 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Thursday 26 March 2026  05:11:55 +0000 (0:00:02.047)       0:09:19.591 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Thursday 26 March 2026  05:11:57 +0000 (0:00:01.416)       0:09:21.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Thursday 26 March 2026  05:11:58 +0000 (0:00:01.144)       0:09:22.152 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:11:59 +0000 (0:00:01.472)       0:09:23.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:12:01 +0000 (0:00:01.125)       0:09:24.750 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:12:02 +0000 (0:00:01.105)       0:09:25.855 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Thursday 26 March 2026  05:12:03 +0000 (0:00:01.470)       0:09:27.326 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Thursday 26 March 2026  05:12:05 +0000 (0:00:02.330)       0:09:29.657 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Thursday 26 March 2026  05:12:08 +0000 (0:00:02.018)       0:09:31.675 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Thursday 26 March 2026  05:12:10 +0000 (0:00:02.419)       0:09:34.095 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Thursday 26 March 2026  05:12:13 +0000 (0:00:03.290)       0:09:37.386 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Thursday 26 March 2026  05:12:15 +0000 (0:00:01.623)       0:09:39.010 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Thursday 26 March 2026  05:12:17 +0000 (0:00:02.355)       0:09:41.365 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Thursday 26 March 2026  05:12:20 +0000 (0:00:02.967)       0:09:44.332 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Thursday 26 March 2026  05:12:21 +0000 (0:00:01.131)       0:09:45.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m",
                                "",
                                "TASK [Start ceph mgr] **********************************************************",
                                "Thursday 26 March 2026  05:12:31 +0000 (0:00:09.703)       0:09:55.167 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:12:33 +0000 (0:00:02.422)       0:09:57.590 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:12:36 +0000 (0:00:02.229)       0:09:59.820 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Non container | waiting for the monitor to join the quorum...] ***********",
                                "Thursday 26 March 2026  05:12:37 +0000 (0:00:01.430)       0:10:01.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Container | waiting for the containerized monitor to join the quorum...] ***",
                                "Thursday 26 March 2026  05:12:38 +0000 (0:00:01.139)       0:10:02.391 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:12:40 +0000 (0:00:02.224)       0:10:04.615 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:12:42 +0000 (0:00:01.124)       0:10:05.740 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:12:43 +0000 (0:00:01.100)       0:10:06.841 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:12:44 +0000 (0:00:01.117)       0:10:07.958 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:12:45 +0000 (0:00:01.131)       0:10:09.089 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:12:46 +0000 (0:00:01.091)       0:10:10.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:12:47 +0000 (0:00:01.121)       0:10:11.303 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mon cluster] ************************************************",
                                "",
                                "TASK [Remove ceph aliases] *****************************************************",
                                "Thursday 26 March 2026  05:12:48 +0000 (0:00:00.939)       0:10:12.242 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Set mon_host_count] ******************************************************",
                                "Thursday 26 March 2026  05:12:49 +0000 (0:00:01.134)       0:10:13.377 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Fail when less than three monitors] **************************************",
                                "Thursday 26 March 2026  05:12:50 +0000 (0:00:00.796)       0:10:14.173 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Select a running monitor] ************************************************",
                                "Thursday 26 March 2026  05:12:51 +0000 (0:00:00.751)       0:10:14.925 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:12:52 +0000 (0:00:00.774)       0:10:15.699 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:12:53 +0000 (0:00:01.197)       0:10:16.897 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:12:54 +0000 (0:00:01.482)       0:10:18.379 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:12:55 +0000 (0:00:01.155)       0:10:19.535 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:12:57 +0000 (0:00:01.444)       0:10:20.979 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:12:58 +0000 (0:00:01.141)       0:10:22.120 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:12:59 +0000 (0:00:01.146)       0:10:23.267 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:13:00 +0000 (0:00:01.138)       0:10:24.406 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:13:01 +0000 (0:00:01.137)       0:10:25.544 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:13:03 +0000 (0:00:01.225)       0:10:26.769 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:13:05 +0000 (0:00:01.979)       0:10:28.749 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:13:06 +0000 (0:00:01.291)       0:10:30.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:13:09 +0000 (0:00:03.220)       0:10:33.261 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:13:11 +0000 (0:00:01.731)       0:10:34.993 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:13:13 +0000 (0:00:01.960)       0:10:36.954 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:13:14 +0000 (0:00:01.173)       0:10:38.127 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:13:06.930490', 'end': '2026-03-26 05:13:06.985292', 'delta': '0:00:00.054802', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': '1fb5a820b9f6', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:13:07.808352', 'end': '2026-03-26 05:13:07.863473', 'delta': '0:00:00.055121', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1fb5a820b9f6'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:13:08.378617', 'end': '2026-03-26 05:13:08.427807', 'delta': '0:00:00.049190', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:13:15 +0000 (0:00:01.248)       0:10:39.375 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:13:16 +0000 (0:00:01.265)       0:10:40.640 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:13:18 +0000 (0:00:01.289)       0:10:41.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:13:19 +0000 (0:00:01.171)       0:10:43.101 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:13:22 +0000 (0:00:02.994)       0:10:46.095 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:13:23 +0000 (0:00:01.136)       0:10:47.232 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:13:24 +0000 (0:00:01.163)       0:10:48.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:13:25 +0000 (0:00:01.199)       0:10:49.595 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:13:27 +0000 (0:00:01.130)       0:10:50.725 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:13:28 +0000 (0:00:01.148)       0:10:51.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:13:29 +0000 (0:00:01.191)       0:10:53.065 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:13:30 +0000 (0:00:01.161)       0:10:54.227 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:13:31 +0000 (0:00:01.180)       0:10:55.408 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:13:32 +0000 (0:00:01.116)       0:10:56.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:13:33 +0000 (0:00:01.105)       0:10:57.630 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:13:35 +0000 (0:00:01.240)       0:10:58.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:13:36 +0000 (0:00:01.281)       0:11:00.152 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:13:38 +0000 (0:00:01.512)       0:11:01.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:13:39 +0000 (0:00:01.166)       0:11:02.831 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:13:40 +0000 (0:00:01.499)       0:11:04.331 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:13:41 +0000 (0:00:01.115)       0:11:05.447 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:13:43 +0000 (0:00:01.408)       0:11:06.855 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:13:44 +0000 (0:00:01.164)       0:11:08.019 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:13:46 +0000 (0:00:02.048)       0:11:10.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:13:47 +0000 (0:00:01.208)       0:11:11.277 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:13:48 +0000 (0:00:01.111)       0:11:12.389 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:13:50 +0000 (0:00:01.883)       0:11:14.272 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get ceph cluster status] *************************************************",
                                "Thursday 26 March 2026  05:13:53 +0000 (0:00:02.391)       0:11:16.664 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Display ceph health detail] **********************************************",
                                "Thursday 26 March 2026  05:13:53 +0000 (0:00:00.872)       0:11:17.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Fail if cluster isn't in an acceptable state] ****************************",
                                "Thursday 26 March 2026  05:13:54 +0000 (0:00:00.876)       0:11:18.414 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Get the ceph quorum status] **********************************************",
                                "Thursday 26 March 2026  05:13:55 +0000 (0:00:00.765)       0:11:19.180 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Fail if the cluster quorum isn't in an acceptable state] *****************",
                                "Thursday 26 March 2026  05:13:56 +0000 (0:00:00.894)       0:11:20.074 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************",
                                "Thursday 26 March 2026  05:13:57 +0000 (0:00:00.785)       0:11:20.859 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Create potentially missing keys (rbd and rbd-mirror)] ********************",
                                "Thursday 26 March 2026  05:13:58 +0000 (0:00:01.087)       0:11:21.947 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-1']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd', 'testbed-node-2']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-1']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=['bootstrap-rbd-mirror', 'testbed-node-2']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Stop ceph mon] ***********************************************************",
                                "Thursday 26 March 2026  05:13:59 +0000 (0:00:01.605)       0:11:23.552 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "",
                                "TASK [Mask the mgr service] ****************************************************",
                                "Thursday 26 March 2026  05:14:03 +0000 (0:00:03.184)       0:11:26.737 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:14:05 +0000 (0:00:01.984)       0:11:28.721 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:14:06 +0000 (0:00:01.137)       0:11:29.859 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:14:07 +0000 (0:00:01.138)       0:11:30.998 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:14:08 +0000 (0:00:01.518)       0:11:32.517 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:14:09 +0000 (0:00:01.128)       0:11:33.645 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:14:11 +0000 (0:00:01.116)       0:11:34.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:14:12 +0000 (0:00:01.238)       0:11:36.000 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:14:13 +0000 (0:00:01.551)       0:11:37.552 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:14:15 +0000 (0:00:01.151)       0:11:38.704 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:14:16 +0000 (0:00:01.162)       0:11:39.866 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:14:17 +0000 (0:00:01.529)       0:11:41.396 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:14:19 +0000 (0:00:01.555)       0:11:42.951 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:14:20 +0000 (0:00:00.806)       0:11:43.758 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:14:20 +0000 (0:00:00.824)       0:11:44.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:14:21 +0000 (0:00:00.801)       0:11:45.383 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:14:22 +0000 (0:00:00.798)       0:11:46.182 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:14:23 +0000 (0:00:00.775)       0:11:46.957 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:14:24 +0000 (0:00:00.805)       0:11:47.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:14:24 +0000 (0:00:00.819)       0:11:48.582 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:14:25 +0000 (0:00:00.865)       0:11:49.448 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:14:26 +0000 (0:00:00.805)       0:11:50.253 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:14:27 +0000 (0:00:00.831)       0:11:51.085 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:14:28 +0000 (0:00:00.766)       0:11:51.852 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:14:28 +0000 (0:00:00.789)       0:11:52.642 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:14:29 +0000 (0:00:00.879)       0:11:53.521 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:14:30 +0000 (0:00:00.760)       0:11:54.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:14:31 +0000 (0:00:00.787)       0:11:55.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:14:32 +0000 (0:00:00.783)       0:11:55.852 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:14:32 +0000 (0:00:00.782)       0:11:56.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:14:33 +0000 (0:00:00.824)       0:11:57.459 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:14:34 +0000 (0:00:00.765)       0:11:58.224 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:14:35 +0000 (0:00:00.833)       0:11:59.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:14:36 +0000 (0:00:00.780)       0:11:59.839 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:14:37 +0000 (0:00:00.826)       0:12:00.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:14:38 +0000 (0:00:01.617)       0:12:02.283 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:14:40 +0000 (0:00:02.084)       0:12:04.367 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:14:41 +0000 (0:00:01.209)       0:12:05.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:14:43 +0000 (0:00:01.134)       0:12:06.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:14:44 +0000 (0:00:01.146)       0:12:07.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:14:46 +0000 (0:00:01.852)       0:12:09.711 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:14:47 +0000 (0:00:01.537)       0:12:11.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:14:48 +0000 (0:00:01.162)       0:12:12.411 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:14:49 +0000 (0:00:00.805)       0:12:13.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:14:50 +0000 (0:00:00.755)       0:12:13.972 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:14:51 +0000 (0:00:01.138)       0:12:15.111 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:14:53 +0000 (0:00:01.856)       0:12:16.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:14:54 +0000 (0:00:01.130)       0:12:18.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:14:55 +0000 (0:00:01.155)       0:12:19.253 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:14:56 +0000 (0:00:01.211)       0:12:20.464 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:14:58 +0000 (0:00:01.200)       0:12:21.665 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:14:59 +0000 (0:00:01.192)       0:12:22.857 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:15:00 +0000 (0:00:00.811)       0:12:23.669 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:15:02 +0000 (0:00:02.213)       0:12:25.883 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:15:03 +0000 (0:00:00.787)       0:12:26.671 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:15:04 +0000 (0:00:01.115)       0:12:27.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:15:05 +0000 (0:00:01.188)       0:12:28.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:15:06 +0000 (0:00:01.138)       0:12:30.114 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:15:07 +0000 (0:00:01.115)       0:12:31.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:15:08 +0000 (0:00:01.155)       0:12:32.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:15:09 +0000 (0:00:01.131)       0:12:33.516 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:15:10 +0000 (0:00:01.140)       0:12:34.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:15:12 +0000 (0:00:01.206)       0:12:35.862 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:15:13 +0000 (0:00:01.119)       0:12:36.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:15:14 +0000 (0:00:00.790)       0:12:37.773 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:15:15 +0000 (0:00:01.126)       0:12:38.899 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:15:21 +0000 (0:00:06.237)       0:12:45.137 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:15:22 +0000 (0:00:00.794)       0:12:45.931 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:15:23 +0000 (0:00:00.780)       0:12:46.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:15:23 +0000 (0:00:00.787)       0:12:47.498 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:15:24 +0000 (0:00:00.796)       0:12:48.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:15:25 +0000 (0:00:00.840)       0:12:49.136 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:15:26 +0000 (0:00:00.814)       0:12:49.951 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:15:27 +0000 (0:00:00.781)       0:12:50.733 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:15:27 +0000 (0:00:00.766)       0:12:51.499 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:15:28 +0000 (0:00:00.783)       0:12:52.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:15:29 +0000 (0:00:00.813)       0:12:53.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:15:30 +0000 (0:00:00.804)       0:12:53.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:15:31 +0000 (0:00:00.800)       0:12:54.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:15:31 +0000 (0:00:00.854)       0:12:55.556 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:15:32 +0000 (0:00:00.833)       0:12:56.390 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:15:33 +0000 (0:00:00.875)       0:12:57.266 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:15:34 +0000 (0:00:00.770)       0:12:58.037 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:15:35 +0000 (0:00:00.773)       0:12:58.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:15:35 +0000 (0:00:00.771)       0:12:59.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:15:36 +0000 (0:00:00.809)       0:13:00.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:15:37 +0000 (0:00:00.774)       0:13:01.167 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:15:38 +0000 (0:00:00.824)       0:13:01.992 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:15:39 +0000 (0:00:01.110)       0:13:03.102 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:15:40 +0000 (0:00:01.063)       0:13:04.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:15:41 +0000 (0:00:01.073)       0:13:05.239 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:15:42 +0000 (0:00:00.768)       0:13:06.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:15:43 +0000 (0:00:00.888)       0:13:06.896 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:15:44 +0000 (0:00:01.395)       0:13:08.292 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Thursday 26 March 2026  05:15:45 +0000 (0:00:00.865)       0:13:09.157 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Thursday 26 March 2026  05:15:46 +0000 (0:00:01.241)       0:13:10.398 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Thursday 26 March 2026  05:15:49 +0000 (0:00:03.163)       0:13:13.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Thursday 26 March 2026  05:15:51 +0000 (0:00:01.211)       0:13:14.774 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Thursday 26 March 2026  05:15:52 +0000 (0:00:01.156)       0:13:15.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Thursday 26 March 2026  05:15:53 +0000 (0:00:01.142)       0:13:17.073 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Thursday 26 March 2026  05:15:55 +0000 (0:00:02.017)       0:13:19.091 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Thursday 26 March 2026  05:15:56 +0000 (0:00:01.554)       0:13:20.646 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Thursday 26 March 2026  05:15:58 +0000 (0:00:01.544)       0:13:22.190 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Thursday 26 March 2026  05:16:00 +0000 (0:00:01.473)       0:13:23.663 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Thursday 26 March 2026  05:16:01 +0000 (0:00:01.616)       0:13:25.281 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Thursday 26 March 2026  05:16:03 +0000 (0:00:01.538)       0:13:26.819 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Thursday 26 March 2026  05:16:07 +0000 (0:00:04.191)       0:13:31.011 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Thursday 26 March 2026  05:16:09 +0000 (0:00:02.026)       0:13:33.038 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Thursday 26 March 2026  05:16:10 +0000 (0:00:01.175)       0:13:34.213 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Thursday 26 March 2026  05:16:11 +0000 (0:00:01.129)       0:13:35.343 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Thursday 26 March 2026  05:16:13 +0000 (0:00:01.741)       0:13:37.085 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Thursday 26 March 2026  05:16:15 +0000 (0:00:01.576)       0:13:38.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Thursday 26 March 2026  05:16:15 +0000 (0:00:00.808)       0:13:39.470 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:16:16 +0000 (0:00:01.167)       0:13:40.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:16:18 +0000 (0:00:01.170)       0:13:41.809 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:16:19 +0000 (0:00:01.165)       0:13:42.974 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Thursday 26 March 2026  05:16:20 +0000 (0:00:01.140)       0:13:44.114 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Thursday 26 March 2026  05:16:23 +0000 (0:00:02.643)       0:13:46.758 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Thursday 26 March 2026  05:16:25 +0000 (0:00:01.972)       0:13:48.730 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Thursday 26 March 2026  05:16:27 +0000 (0:00:02.330)       0:13:51.061 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Thursday 26 March 2026  05:16:30 +0000 (0:00:03.186)       0:13:54.247 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Thursday 26 March 2026  05:16:31 +0000 (0:00:01.118)       0:13:55.366 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Thursday 26 March 2026  05:16:54 +0000 (0:00:22.816)       0:14:18.182 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Thursday 26 March 2026  05:16:57 +0000 (0:00:02.677)       0:14:20.859 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Thursday 26 March 2026  05:16:58 +0000 (0:00:00.807)       0:14:21.667 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m",
                                "",
                                "TASK [Start ceph mgr] **********************************************************",
                                "Thursday 26 March 2026  05:17:07 +0000 (0:00:09.327)       0:14:30.995 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:17:09 +0000 (0:00:02.117)       0:14:33.112 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:17:11 +0000 (0:00:01.877)       0:14:34.989 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Non container | waiting for the monitor to join the quorum...] ***********",
                                "Thursday 26 March 2026  05:17:12 +0000 (0:00:01.109)       0:14:36.099 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Container | waiting for the containerized monitor to join the quorum...] ***",
                                "Thursday 26 March 2026  05:17:13 +0000 (0:00:00.761)       0:14:36.861 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:17:15 +0000 (0:00:02.347)       0:14:39.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:17:16 +0000 (0:00:00.800)       0:14:40.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:17:17 +0000 (0:00:00.794)       0:14:40.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:17:17 +0000 (0:00:00.773)       0:14:41.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:17:18 +0000 (0:00:00.798)       0:14:42.375 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:17:19 +0000 (0:00:00.784)       0:14:43.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:17:20 +0000 (0:00:00.800)       0:14:43.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mon cluster] ************************************************",
                                "",
                                "TASK [Remove ceph aliases] *****************************************************",
                                "Thursday 26 March 2026  05:17:21 +0000 (0:00:00.955)       0:14:44.915 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set mon_host_count] ******************************************************",
                                "Thursday 26 March 2026  05:17:22 +0000 (0:00:01.157)       0:14:46.072 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Fail when less than three monitors] **************************************",
                                "Thursday 26 March 2026  05:17:23 +0000 (0:00:00.794)       0:14:46.867 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Select a running monitor] ************************************************",
                                "Thursday 26 March 2026  05:17:23 +0000 (0:00:00.779)       0:14:47.647 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:17:24 +0000 (0:00:00.820)       0:14:48.468 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:17:25 +0000 (0:00:01.154)       0:14:49.622 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:17:27 +0000 (0:00:01.440)       0:14:51.063 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:17:28 +0000 (0:00:01.110)       0:14:52.173 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:17:29 +0000 (0:00:01.443)       0:14:53.617 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:17:31 +0000 (0:00:01.175)       0:14:54.793 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:17:32 +0000 (0:00:01.169)       0:14:55.963 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:17:33 +0000 (0:00:01.280)       0:14:57.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:17:34 +0000 (0:00:01.231)       0:14:58.476 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:17:35 +0000 (0:00:01.169)       0:14:59.645 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:17:38 +0000 (0:00:02.218)       0:15:01.863 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:17:39 +0000 (0:00:01.300)       0:15:03.164 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:17:42 +0000 (0:00:03.182)       0:15:06.346 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:17:44 +0000 (0:00:01.424)       0:15:07.771 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:17:45 +0000 (0:00:01.660)       0:15:09.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:17:46 +0000 (0:00:01.154)       0:15:10.586 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:17:40.373487', 'end': '2026-03-26 05:17:40.412891', 'delta': '0:00:00.039404', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:17:40.900551', 'end': '2026-03-26 05:17:40.944725', 'delta': '0:00:00.044174', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': '2a382ea60872', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:17:41.454921', 'end': '2026-03-26 05:17:41.516677', 'delta': '0:00:00.061756', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['2a382ea60872'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:17:48 +0000 (0:00:01.214)       0:15:11.801 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:17:49 +0000 (0:00:01.310)       0:15:13.111 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:17:50 +0000 (0:00:01.285)       0:15:14.397 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:17:51 +0000 (0:00:01.202)       0:15:15.600 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:17:53 +0000 (0:00:02.005)       0:15:17.605 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:17:55 +0000 (0:00:01.157)       0:15:18.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:17:56 +0000 (0:00:01.150)       0:15:19.914 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:17:57 +0000 (0:00:01.207)       0:15:21.122 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:17:58 +0000 (0:00:01.132)       0:15:22.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:17:59 +0000 (0:00:01.178)       0:15:23.433 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:18:00 +0000 (0:00:01.137)       0:15:24.571 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:18:02 +0000 (0:00:01.158)       0:15:25.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:18:03 +0000 (0:00:01.142)       0:15:26.872 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:18:04 +0000 (0:00:01.146)       0:15:28.019 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:18:05 +0000 (0:00:01.140)       0:15:29.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:18:06 +0000 (0:00:01.292)       0:15:30.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:18:08 +0000 (0:00:01.222)       0:15:31.675 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:18:09 +0000 (0:00:01.485)       0:15:33.164 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:18:10 +0000 (0:00:01.164)       0:15:34.328 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:18:12 +0000 (0:00:01.508)       0:15:35.837 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:18:13 +0000 (0:00:01.109)       0:15:36.947 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:18:14 +0000 (0:00:01.278)       0:15:38.225 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:18:15 +0000 (0:00:01.176)       0:15:39.402 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:18:17 +0000 (0:00:01.686)       0:15:41.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:18:18 +0000 (0:00:01.213)       0:15:42.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:18:19 +0000 (0:00:01.119)       0:15:43.421 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:18:21 +0000 (0:00:01.904)       0:15:45.326 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get ceph cluster status] *************************************************",
                                "Thursday 26 March 2026  05:18:23 +0000 (0:00:02.305)       0:15:47.632 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Display ceph health detail] **********************************************",
                                "Thursday 26 March 2026  05:18:24 +0000 (0:00:00.919)       0:15:48.551 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Fail if cluster isn't in an acceptable state] ****************************",
                                "Thursday 26 March 2026  05:18:25 +0000 (0:00:00.932)       0:15:49.484 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get the ceph quorum status] **********************************************",
                                "Thursday 26 March 2026  05:18:26 +0000 (0:00:00.789)       0:15:50.274 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Fail if the cluster quorum isn't in an acceptable state] *****************",
                                "Thursday 26 March 2026  05:18:27 +0000 (0:00:00.916)       0:15:51.191 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Ensure /var/lib/ceph/bootstrap-rbd-mirror is present] ********************",
                                "Thursday 26 March 2026  05:18:28 +0000 (0:00:00.769)       0:15:51.961 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Create potentially missing keys (rbd and rbd-mirror)] ********************",
                                "Thursday 26 March 2026  05:18:29 +0000 (0:00:01.408)       0:15:53.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-1']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd', 'testbed-node-2']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-1']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=['bootstrap-rbd-mirror', 'testbed-node-2']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Stop ceph mon] ***********************************************************",
                                "Thursday 26 March 2026  05:18:31 +0000 (0:00:01.623)       0:15:54.993 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [Mask the mgr service] ****************************************************",
                                "Thursday 26 March 2026  05:18:34 +0000 (0:00:03.431)       0:15:58.424 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:18:37 +0000 (0:00:02.252)       0:16:00.677 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:18:38 +0000 (0:00:01.081)       0:16:01.758 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:18:39 +0000 (0:00:01.146)       0:16:02.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:18:40 +0000 (0:00:01.594)       0:16:04.499 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:18:41 +0000 (0:00:01.137)       0:16:05.636 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:18:43 +0000 (0:00:01.137)       0:16:06.774 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:18:44 +0000 (0:00:01.107)       0:16:07.881 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:18:45 +0000 (0:00:01.586)       0:16:09.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:18:46 +0000 (0:00:01.119)       0:16:10.588 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:18:48 +0000 (0:00:01.187)       0:16:11.776 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:18:49 +0000 (0:00:01.609)       0:16:13.386 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:18:51 +0000 (0:00:01.563)       0:16:14.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:18:52 +0000 (0:00:00.777)       0:16:15.728 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:18:52 +0000 (0:00:00.767)       0:16:16.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:18:53 +0000 (0:00:00.800)       0:16:17.296 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:18:54 +0000 (0:00:00.802)       0:16:18.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:18:55 +0000 (0:00:00.789)       0:16:18.888 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:18:56 +0000 (0:00:00.775)       0:16:19.664 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:18:56 +0000 (0:00:00.779)       0:16:20.443 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:18:57 +0000 (0:00:00.810)       0:16:21.254 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:18:58 +0000 (0:00:00.800)       0:16:22.054 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:18:59 +0000 (0:00:00.792)       0:16:22.846 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:19:00 +0000 (0:00:00.822)       0:16:23.669 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:19:00 +0000 (0:00:00.769)       0:16:24.439 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:19:01 +0000 (0:00:00.769)       0:16:25.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:19:02 +0000 (0:00:00.799)       0:16:26.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:19:03 +0000 (0:00:00.755)       0:16:26.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:19:03 +0000 (0:00:00.746)       0:16:27.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:19:04 +0000 (0:00:00.769)       0:16:28.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:19:05 +0000 (0:00:00.778)       0:16:29.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:19:06 +0000 (0:00:00.765)       0:16:29.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:19:06 +0000 (0:00:00.758)       0:16:30.583 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:19:07 +0000 (0:00:00.757)       0:16:31.341 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:19:08 +0000 (0:00:00.767)       0:16:32.109 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:19:10 +0000 (0:00:01.645)       0:16:33.755 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:19:12 +0000 (0:00:02.045)       0:16:35.800 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:19:13 +0000 (0:00:01.119)       0:16:36.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:19:14 +0000 (0:00:01.130)       0:16:38.051 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:19:15 +0000 (0:00:01.163)       0:16:39.215 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:19:17 +0000 (0:00:01.897)       0:16:41.112 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:19:18 +0000 (0:00:01.455)       0:16:42.568 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:19:20 +0000 (0:00:01.125)       0:16:43.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:19:20 +0000 (0:00:00.803)       0:16:44.497 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:19:21 +0000 (0:00:00.774)       0:16:45.272 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:19:22 +0000 (0:00:01.137)       0:16:46.410 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:19:24 +0000 (0:00:01.892)       0:16:48.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:19:25 +0000 (0:00:01.145)       0:16:49.448 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:19:26 +0000 (0:00:01.136)       0:16:50.584 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:19:28 +0000 (0:00:01.197)       0:16:51.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:19:29 +0000 (0:00:01.157)       0:16:52.939 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:19:30 +0000 (0:00:01.179)       0:16:54.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:19:31 +0000 (0:00:00.818)       0:16:54.937 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:19:33 +0000 (0:00:02.279)       0:16:57.216 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:19:34 +0000 (0:00:00.763)       0:16:57.980 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:19:35 +0000 (0:00:01.138)       0:16:59.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:19:36 +0000 (0:00:01.132)       0:17:00.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:19:37 +0000 (0:00:01.140)       0:17:01.391 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:19:38 +0000 (0:00:01.184)       0:17:02.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:19:40 +0000 (0:00:01.154)       0:17:03.730 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:19:41 +0000 (0:00:01.219)       0:17:04.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:19:42 +0000 (0:00:01.154)       0:17:06.105 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:19:43 +0000 (0:00:01.176)       0:17:07.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:19:44 +0000 (0:00:01.206)       0:17:08.487 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:19:45 +0000 (0:00:00.849)       0:17:09.337 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:19:46 +0000 (0:00:01.103)       0:17:10.441 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:19:53 +0000 (0:00:06.354)       0:17:16.795 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:19:53 +0000 (0:00:00.763)       0:17:17.558 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:19:54 +0000 (0:00:00.774)       0:17:18.333 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:19:55 +0000 (0:00:00.800)       0:17:19.134 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:19:56 +0000 (0:00:00.817)       0:17:19.952 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:19:57 +0000 (0:00:00.834)       0:17:20.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:19:58 +0000 (0:00:00.894)       0:17:21.680 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:19:58 +0000 (0:00:00.847)       0:17:22.528 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:19:59 +0000 (0:00:00.773)       0:17:23.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:20:00 +0000 (0:00:00.800)       0:17:24.102 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:20:01 +0000 (0:00:00.767)       0:17:24.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:20:01 +0000 (0:00:00.777)       0:17:25.648 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:20:02 +0000 (0:00:00.786)       0:17:26.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:20:03 +0000 (0:00:00.894)       0:17:27.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:20:04 +0000 (0:00:00.770)       0:17:28.099 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:20:05 +0000 (0:00:00.882)       0:17:28.981 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:20:06 +0000 (0:00:00.758)       0:17:29.740 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:20:06 +0000 (0:00:00.784)       0:17:30.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:20:07 +0000 (0:00:00.743)       0:17:31.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:20:08 +0000 (0:00:00.780)       0:17:32.049 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:20:09 +0000 (0:00:00.762)       0:17:32.812 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:20:09 +0000 (0:00:00.748)       0:17:33.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:20:10 +0000 (0:00:01.070)       0:17:34.631 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:20:12 +0000 (0:00:01.039)       0:17:35.670 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:20:13 +0000 (0:00:01.016)       0:17:36.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:20:13 +0000 (0:00:00.772)       0:17:37.459 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:20:14 +0000 (0:00:00.929)       0:17:38.389 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:20:16 +0000 (0:00:01.440)       0:17:39.830 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Thursday 26 March 2026  05:20:17 +0000 (0:00:00.829)       0:17:40.660 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Thursday 26 March 2026  05:20:18 +0000 (0:00:01.164)       0:17:41.824 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Thursday 26 March 2026  05:20:21 +0000 (0:00:03.209)       0:17:45.033 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Thursday 26 March 2026  05:20:22 +0000 (0:00:01.188)       0:17:46.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Thursday 26 March 2026  05:20:23 +0000 (0:00:01.153)       0:17:47.375 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Thursday 26 March 2026  05:20:24 +0000 (0:00:01.164)       0:17:48.540 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Thursday 26 March 2026  05:20:26 +0000 (0:00:02.091)       0:17:50.631 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Thursday 26 March 2026  05:20:28 +0000 (0:00:01.605)       0:17:52.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Thursday 26 March 2026  05:20:30 +0000 (0:00:01.522)       0:17:53.759 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Thursday 26 March 2026  05:20:31 +0000 (0:00:01.533)       0:17:55.293 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Thursday 26 March 2026  05:20:33 +0000 (0:00:01.584)       0:17:56.877 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Thursday 26 March 2026  05:20:34 +0000 (0:00:01.548)       0:17:58.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Thursday 26 March 2026  05:20:38 +0000 (0:00:04.195)       0:18:02.621 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Thursday 26 March 2026  05:20:41 +0000 (0:00:02.120)       0:18:04.742 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Thursday 26 March 2026  05:20:42 +0000 (0:00:01.146)       0:18:05.888 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Thursday 26 March 2026  05:20:43 +0000 (0:00:01.151)       0:18:07.039 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Thursday 26 March 2026  05:20:45 +0000 (0:00:01.799)       0:18:08.838 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Thursday 26 March 2026  05:20:46 +0000 (0:00:01.465)       0:18:10.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Thursday 26 March 2026  05:20:47 +0000 (0:00:00.760)       0:18:11.065 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:20:48 +0000 (0:00:01.123)       0:18:12.189 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:20:49 +0000 (0:00:01.118)       0:18:13.308 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:20:50 +0000 (0:00:01.176)       0:18:14.484 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Thursday 26 March 2026  05:20:51 +0000 (0:00:01.095)       0:18:15.579 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Thursday 26 March 2026  05:20:54 +0000 (0:00:02.713)       0:18:18.292 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Thursday 26 March 2026  05:20:56 +0000 (0:00:02.065)       0:18:20.358 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Thursday 26 March 2026  05:20:59 +0000 (0:00:02.468)       0:18:22.827 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Thursday 26 March 2026  05:21:03 +0000 (0:00:04.019)       0:18:26.846 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Thursday 26 March 2026  05:21:04 +0000 (0:00:01.151)       0:18:27.998 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Thursday 26 March 2026  05:21:27 +0000 (0:00:23.007)       0:18:51.006 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Thursday 26 March 2026  05:21:30 +0000 (0:00:02.689)       0:18:53.695 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Thursday 26 March 2026  05:21:30 +0000 (0:00:00.801)       0:18:54.496 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__77be735a24b84ab013e2e181233ee69b9d9f8b69'}]) \u001b[0m",
                                "",
                                "TASK [Start ceph mgr] **********************************************************",
                                "Thursday 26 March 2026  05:21:40 +0000 (0:00:09.632)       0:19:04.129 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:21:42 +0000 (0:00:02.183)       0:19:06.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:21:44 +0000 (0:00:01.868)       0:19:08.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Non container | waiting for the monitor to join the quorum...] ***********",
                                "Thursday 26 March 2026  05:21:45 +0000 (0:00:01.420)       0:19:09.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Container | waiting for the containerized monitor to join the quorum...] ***",
                                "Thursday 26 March 2026  05:21:46 +0000 (0:00:00.762)       0:19:10.364 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:21:48 +0000 (0:00:01.963)       0:19:12.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:21:49 +0000 (0:00:00.825)       0:19:13.154 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:21:50 +0000 (0:00:00.762)       0:19:13.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:21:51 +0000 (0:00:00.773)       0:19:14.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:21:51 +0000 (0:00:00.754)       0:19:15.444 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:21:52 +0000 (0:00:00.764)       0:19:16.209 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:21:53 +0000 (0:00:00.845)       0:19:17.055 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reset mon_host] **********************************************************",
                                "",
                                "TASK [Reset mon_host fact] *****************************************************",
                                "Thursday 26 March 2026  05:21:55 +0000 (0:00:01.781)       0:19:18.836 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:21:56 +0000 (0:00:01.460)       0:19:20.297 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:21:57 +0000 (0:00:01.175)       0:19:21.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:21:58 +0000 (0:00:01.170)       0:19:22.644 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:22:00 +0000 (0:00:01.141)       0:19:23.785 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:22:01 +0000 (0:00:01.174)       0:19:24.959 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:22:02 +0000 (0:00:01.123)       0:19:26.083 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:22:03 +0000 (0:00:01.125)       0:19:27.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:22:04 +0000 (0:00:01.125)       0:19:28.334 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:22:05 +0000 (0:00:01.134)       0:19:29.469 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:22:06 +0000 (0:00:01.108)       0:19:30.578 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:22:08 +0000 (0:00:01.120)       0:19:31.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:22:09 +0000 (0:00:01.107)       0:19:32.806 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:22:10 +0000 (0:00:01.133)       0:19:33.939 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:22:11 +0000 (0:00:01.194)       0:19:35.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:22:12 +0000 (0:00:01.154)       0:19:36.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:22:13 +0000 (0:00:01.114)       0:19:37.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:22:14 +0000 (0:00:01.123)       0:19:38.526 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:22:15 +0000 (0:00:01.121)       0:19:39.648 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:22:17 +0000 (0:00:01.103)       0:19:40.751 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:22:18 +0000 (0:00:01.113)       0:19:41.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:22:19 +0000 (0:00:01.172)       0:19:43.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:22:20 +0000 (0:00:01.113)       0:19:44.151 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:22:21 +0000 (0:00:01.187)       0:19:45.339 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:22:22 +0000 (0:00:01.154)       0:19:46.494 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:22:23 +0000 (0:00:01.131)       0:19:47.625 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:22:25 +0000 (0:00:01.147)       0:19:48.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:22:26 +0000 (0:00:01.146)       0:19:49.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:22:27 +0000 (0:00:01.144)       0:19:51.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:22:28 +0000 (0:00:01.208)       0:19:52.273 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:22:29 +0000 (0:00:01.150)       0:19:53.424 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:22:30 +0000 (0:00:01.177)       0:19:54.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:22:32 +0000 (0:00:01.125)       0:19:55.728 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:22:33 +0000 (0:00:01.243)       0:19:56.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:22:34 +0000 (0:00:01.092)       0:19:58.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:22:35 +0000 (0:00:01.138)       0:19:59.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:22:36 +0000 (0:00:01.115)       0:20:00.318 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:22:37 +0000 (0:00:01.113)       0:20:01.432 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:22:38 +0000 (0:00:01.132)       0:20:02.564 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:22:40 +0000 (0:00:01.128)       0:20:03.692 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:22:41 +0000 (0:00:01.147)       0:20:04.840 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:22:42 +0000 (0:00:01.204)       0:20:06.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:22:43 +0000 (0:00:01.141)       0:20:07.187 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:22:44 +0000 (0:00:01.124)       0:20:08.312 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:22:45 +0000 (0:00:01.109)       0:20:09.422 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:22:46 +0000 (0:00:01.096)       0:20:10.518 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:22:48 +0000 (0:00:01.165)       0:20:11.684 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:22:49 +0000 (0:00:01.245)       0:20:12.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:22:50 +0000 (0:00:01.181)       0:20:14.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:22:51 +0000 (0:00:01.283)       0:20:15.393 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:22:52 +0000 (0:00:01.180)       0:20:16.574 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:22:54 +0000 (0:00:01.124)       0:20:17.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:22:55 +0000 (0:00:01.155)       0:20:18.853 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:22:56 +0000 (0:00:01.160)       0:20:20.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:22:57 +0000 (0:00:01.128)       0:20:21.143 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:22:58 +0000 (0:00:01.155)       0:20:22.298 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:23:00 +0000 (0:00:01.393)       0:20:23.692 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:23:01 +0000 (0:00:01.760)       0:20:25.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:23:03 +0000 (0:00:01.710)       0:20:27.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:23:04 +0000 (0:00:01.196)       0:20:28.359 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:23:05 +0000 (0:00:01.236)       0:20:29.596 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:23:07 +0000 (0:00:01.150)       0:20:30.746 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:23:08 +0000 (0:00:01.401)       0:20:32.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:23:09 +0000 (0:00:01.170)       0:20:33.318 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:23:10 +0000 (0:00:01.141)       0:20:34.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:23:11 +0000 (0:00:01.169)       0:20:35.630 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:23:12 +0000 (0:00:00.967)       0:20:36.598 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:23:13 +0000 (0:00:00.781)       0:20:37.380 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:23:14 +0000 (0:00:00.869)       0:20:38.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:23:15 +0000 (0:00:00.820)       0:20:39.070 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:23:16 +0000 (0:00:00.771)       0:20:39.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:23:16 +0000 (0:00:00.811)       0:20:40.653 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:23:17 +0000 (0:00:00.792)       0:20:41.446 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:23:18 +0000 (0:00:00.772)       0:20:42.218 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:23:19 +0000 (0:00:00.782)       0:20:43.001 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:23:20 +0000 (0:00:00.798)       0:20:43.799 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:23:20 +0000 (0:00:00.797)       0:20:44.597 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:23:21 +0000 (0:00:00.806)       0:20:45.403 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:23:22 +0000 (0:00:00.787)       0:20:46.191 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:23:23 +0000 (0:00:00.808)       0:20:47.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:23:24 +0000 (0:00:00.810)       0:20:47.810 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:23:24 +0000 (0:00:00.780)       0:20:48.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:23:25 +0000 (0:00:00.798)       0:20:49.389 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:23:26 +0000 (0:00:00.784)       0:20:50.174 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:23:27 +0000 (0:00:00.815)       0:20:50.989 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:23:28 +0000 (0:00:00.775)       0:20:51.765 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:23:28 +0000 (0:00:00.785)       0:20:52.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:23:29 +0000 (0:00:00.793)       0:20:53.343 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:23:30 +0000 (0:00:00.771)       0:20:54.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:23:31 +0000 (0:00:00.776)       0:20:54.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:23:32 +0000 (0:00:00.767)       0:20:55.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:23:32 +0000 (0:00:00.799)       0:20:56.459 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:23:33 +0000 (0:00:00.835)       0:20:57.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:23:34 +0000 (0:00:00.813)       0:20:58.108 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:23:35 +0000 (0:00:00.786)       0:20:58.895 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:23:36 +0000 (0:00:00.797)       0:20:59.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:23:36 +0000 (0:00:00.758)       0:21:00.451 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:23:37 +0000 (0:00:00.782)       0:21:01.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:23:38 +0000 (0:00:00.824)       0:21:02.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:23:39 +0000 (0:00:00.784)       0:21:02.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:23:39 +0000 (0:00:00.767)       0:21:03.610 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:23:40 +0000 (0:00:00.867)       0:21:04.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:23:41 +0000 (0:00:00.785)       0:21:05.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:23:42 +0000 (0:00:00.801)       0:21:06.065 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:23:43 +0000 (0:00:00.780)       0:21:06.845 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:23:43 +0000 (0:00:00.780)       0:21:07.625 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:23:44 +0000 (0:00:00.836)       0:21:08.462 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:23:45 +0000 (0:00:00.786)       0:21:09.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:23:46 +0000 (0:00:00.763)       0:21:10.012 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:23:47 +0000 (0:00:00.797)       0:21:10.810 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:23:47 +0000 (0:00:00.761)       0:21:11.572 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:23:48 +0000 (0:00:00.781)       0:21:12.353 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:23:49 +0000 (0:00:00.898)       0:21:13.252 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:23:50 +0000 (0:00:00.781)       0:21:14.033 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:23:51 +0000 (0:00:00.885)       0:21:14.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:23:52 +0000 (0:00:00.816)       0:21:15.735 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:23:52 +0000 (0:00:00.775)       0:21:16.511 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:23:53 +0000 (0:00:00.787)       0:21:17.298 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:23:54 +0000 (0:00:00.806)       0:21:18.104 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:23:55 +0000 (0:00:00.815)       0:21:18.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:23:56 +0000 (0:00:00.816)       0:21:19.737 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:23:57 +0000 (0:00:01.025)       0:21:20.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:23:58 +0000 (0:00:01.084)       0:21:21.846 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:23:59 +0000 (0:00:01.080)       0:21:22.927 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:24:00 +0000 (0:00:00.766)       0:21:23.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:24:00 +0000 (0:00:00.911)       0:21:24.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:24:01 +0000 (0:00:00.804)       0:21:25.410 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:24:03 +0000 (0:00:01.438)       0:21:26.848 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:24:03 +0000 (0:00:00.788)       0:21:27.637 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:24:04 +0000 (0:00:00.815)       0:21:28.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:24:05 +0000 (0:00:00.790)       0:21:29.243 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes when implicitly collocated on monitors] ***********",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:24:06 +0000 (0:00:01.318)       0:21:30.561 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:24:07 +0000 (0:00:00.785)       0:21:31.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:24:08 +0000 (0:00:00.769)       0:21:32.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:24:09 +0000 (0:00:00.788)       0:21:32.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:24:10 +0000 (0:00:00.760)       0:21:33.665 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:24:10 +0000 (0:00:00.800)       0:21:34.466 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:24:11 +0000 (0:00:00.767)       0:21:35.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:24:12 +0000 (0:00:00.764)       0:21:35.998 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:24:13 +0000 (0:00:00.774)       0:21:36.772 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:24:13 +0000 (0:00:00.810)       0:21:37.583 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:24:14 +0000 (0:00:00.778)       0:21:38.362 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:24:15 +0000 (0:00:00.781)       0:21:39.144 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:24:16 +0000 (0:00:00.772)       0:21:39.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:24:17 +0000 (0:00:00.838)       0:21:40.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:24:17 +0000 (0:00:00.785)       0:21:41.540 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:24:18 +0000 (0:00:00.772)       0:21:42.313 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:24:19 +0000 (0:00:00.789)       0:21:43.102 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:24:20 +0000 (0:00:00.770)       0:21:43.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:24:20 +0000 (0:00:00.758)       0:21:44.632 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:24:21 +0000 (0:00:00.793)       0:21:45.425 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:24:22 +0000 (0:00:00.833)       0:21:46.258 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:24:23 +0000 (0:00:00.781)       0:21:47.040 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:24:24 +0000 (0:00:00.810)       0:21:47.851 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:24:24 +0000 (0:00:00.778)       0:21:48.630 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:24:25 +0000 (0:00:00.769)       0:21:49.399 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:24:26 +0000 (0:00:00.815)       0:21:50.215 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:24:27 +0000 (0:00:00.807)       0:21:51.023 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:24:28 +0000 (0:00:00.766)       0:21:51.789 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:24:28 +0000 (0:00:00.771)       0:21:52.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:24:29 +0000 (0:00:00.755)       0:21:53.316 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:24:30 +0000 (0:00:00.756)       0:21:54.072 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:24:31 +0000 (0:00:00.767)       0:21:54.840 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:24:32 +0000 (0:00:00.897)       0:21:55.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:24:32 +0000 (0:00:00.824)       0:21:56.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:24:33 +0000 (0:00:00.821)       0:21:57.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:24:34 +0000 (0:00:00.776)       0:21:58.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:24:35 +0000 (0:00:00.772)       0:21:58.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:24:36 +0000 (0:00:00.775)       0:21:59.710 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:24:36 +0000 (0:00:00.794)       0:22:00.505 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:24:37 +0000 (0:00:00.774)       0:22:01.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:24:38 +0000 (0:00:00.774)       0:22:02.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:24:39 +0000 (0:00:00.806)       0:22:02.860 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:24:39 +0000 (0:00:00.757)       0:22:03.617 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:24:40 +0000 (0:00:00.791)       0:22:04.409 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:24:41 +0000 (0:00:00.815)       0:22:05.225 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:24:42 +0000 (0:00:00.811)       0:22:06.036 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:24:43 +0000 (0:00:00.899)       0:22:06.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:24:44 +0000 (0:00:00.782)       0:22:07.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:24:44 +0000 (0:00:00.858)       0:22:08.577 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:24:45 +0000 (0:00:00.774)       0:22:09.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:24:46 +0000 (0:00:00.814)       0:22:10.166 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:24:47 +0000 (0:00:00.837)       0:22:11.004 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:24:48 +0000 (0:00:00.763)       0:22:11.767 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:24:48 +0000 (0:00:00.787)       0:22:12.555 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:24:49 +0000 (0:00:00.774)       0:22:13.330 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:24:50 +0000 (0:00:01.089)       0:22:14.419 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:24:52 +0000 (0:00:01.469)       0:22:15.889 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:24:53 +0000 (0:00:01.376)       0:22:17.266 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:24:54 +0000 (0:00:00.795)       0:22:18.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:24:55 +0000 (0:00:00.901)       0:22:18.963 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:24:56 +0000 (0:00:00.795)       0:22:19.758 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:24:57 +0000 (0:00:01.055)       0:22:20.814 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:24:57 +0000 (0:00:00.778)       0:22:21.593 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:24:58 +0000 (0:00:00.766)       0:22:22.360 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:24:59 +0000 (0:00:00.809)       0:22:23.170 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes] **************************************************",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:25:00 +0000 (0:00:01.409)       0:22:24.579 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Mask ceph mgr systemd unit] **********************************************",
                                "Thursday 26 March 2026  05:25:13 +0000 (0:00:12.895)       0:22:37.474 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:25:16 +0000 (0:00:02.590)       0:22:40.065 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:25:17 +0000 (0:00:01.141)       0:22:41.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:25:19 +0000 (0:00:01.524)       0:22:42.731 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:25:20 +0000 (0:00:01.146)       0:22:43.877 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:25:21 +0000 (0:00:01.532)       0:22:45.410 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:25:22 +0000 (0:00:01.118)       0:22:46.528 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:25:24 +0000 (0:00:01.136)       0:22:47.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:25:25 +0000 (0:00:01.154)       0:22:48.820 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:25:26 +0000 (0:00:01.157)       0:22:49.977 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:25:27 +0000 (0:00:01.099)       0:22:51.077 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:25:29 +0000 (0:00:01.962)       0:22:53.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:25:30 +0000 (0:00:01.203)       0:22:54.244 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:25:33 +0000 (0:00:03.180)       0:22:57.424 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:25:35 +0000 (0:00:01.779)       0:22:59.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:25:37 +0000 (0:00:01.665)       0:23:00.869 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:25:38 +0000 (0:00:01.192)       0:23:02.061 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:25:31.452374', 'end': '2026-03-26 05:25:31.504874', 'delta': '0:00:00.052500', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:25:32.004082', 'end': '2026-03-26 05:25:32.055136', 'delta': '0:00:00.051054', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:25:32.553894', 'end': '2026-03-26 05:25:32.609169', 'delta': '0:00:00.055275', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:25:39 +0000 (0:00:01.241)       0:23:03.303 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:25:40 +0000 (0:00:01.309)       0:23:04.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:25:42 +0000 (0:00:01.286)       0:23:05.898 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:25:43 +0000 (0:00:01.138)       0:23:07.037 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:25:45 +0000 (0:00:02.009)       0:23:09.046 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:25:46 +0000 (0:00:01.180)       0:23:10.226 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:25:47 +0000 (0:00:01.129)       0:23:11.356 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:25:48 +0000 (0:00:01.190)       0:23:12.547 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:25:50 +0000 (0:00:01.181)       0:23:13.728 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:25:51 +0000 (0:00:01.150)       0:23:14.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:25:52 +0000 (0:00:01.189)       0:23:16.069 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:25:53 +0000 (0:00:01.160)       0:23:17.229 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:25:54 +0000 (0:00:01.140)       0:23:18.370 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:25:55 +0000 (0:00:01.182)       0:23:19.553 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:25:57 +0000 (0:00:01.126)       0:23:20.679 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:25:58 +0000 (0:00:01.346)       0:23:22.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:25:59 +0000 (0:00:01.217)       0:23:23.244 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:26:01 +0000 (0:00:01.553)       0:23:24.797 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:26:02 +0000 (0:00:01.183)       0:23:25.980 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:26:03 +0000 (0:00:01.524)       0:23:27.505 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:26:04 +0000 (0:00:01.131)       0:23:28.636 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:26:06 +0000 (0:00:01.244)       0:23:29.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:26:07 +0000 (0:00:01.138)       0:23:31.019 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:26:09 +0000 (0:00:02.118)       0:23:33.138 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:26:10 +0000 (0:00:01.158)       0:23:34.296 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:26:11 +0000 (0:00:01.167)       0:23:35.463 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:26:13 +0000 (0:00:01.837)       0:23:37.300 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:26:16 +0000 (0:00:02.588)       0:23:39.889 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:26:17 +0000 (0:00:01.119)       0:23:41.009 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:26:18 +0000 (0:00:01.097)       0:23:42.107 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:26:20 +0000 (0:00:01.558)       0:23:43.665 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:26:21 +0000 (0:00:01.112)       0:23:44.778 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:26:22 +0000 (0:00:01.139)       0:23:45.918 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:26:23 +0000 (0:00:01.106)       0:23:47.024 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:26:24 +0000 (0:00:01.584)       0:23:48.609 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:26:26 +0000 (0:00:01.162)       0:23:49.772 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:26:27 +0000 (0:00:01.154)       0:23:50.927 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:26:28 +0000 (0:00:01.595)       0:23:52.522 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:26:30 +0000 (0:00:01.566)       0:23:54.089 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:26:31 +0000 (0:00:01.126)       0:23:55.216 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:26:32 +0000 (0:00:01.179)       0:23:56.396 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:26:33 +0000 (0:00:01.135)       0:23:57.531 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:26:34 +0000 (0:00:01.113)       0:23:58.644 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:26:36 +0000 (0:00:01.131)       0:23:59.776 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:26:37 +0000 (0:00:01.113)       0:24:00.890 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:26:38 +0000 (0:00:01.143)       0:24:02.033 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:26:39 +0000 (0:00:01.229)       0:24:03.263 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:26:40 +0000 (0:00:01.177)       0:24:04.441 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:26:41 +0000 (0:00:01.178)       0:24:05.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:26:43 +0000 (0:00:01.167)       0:24:06.788 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:26:44 +0000 (0:00:01.177)       0:24:07.965 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:26:45 +0000 (0:00:01.153)       0:24:09.119 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:26:46 +0000 (0:00:01.126)       0:24:10.245 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:26:47 +0000 (0:00:01.145)       0:24:11.390 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:26:48 +0000 (0:00:01.144)       0:24:12.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:26:50 +0000 (0:00:01.167)       0:24:13.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:26:51 +0000 (0:00:01.168)       0:24:14.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:26:52 +0000 (0:00:01.152)       0:24:16.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:26:53 +0000 (0:00:01.171)       0:24:17.195 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:26:54 +0000 (0:00:01.142)       0:24:18.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:26:55 +0000 (0:00:01.128)       0:24:19.466 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:26:57 +0000 (0:00:01.990)       0:24:21.457 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:27:00 +0000 (0:00:02.349)       0:24:23.807 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:27:01 +0000 (0:00:01.169)       0:24:24.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:27:02 +0000 (0:00:01.153)       0:24:26.129 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:27:03 +0000 (0:00:01.119)       0:24:27.249 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:27:05 +0000 (0:00:01.818)       0:24:29.067 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:27:06 +0000 (0:00:01.522)       0:24:30.590 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:27:08 +0000 (0:00:01.244)       0:24:31.834 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:27:09 +0000 (0:00:01.141)       0:24:32.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:27:10 +0000 (0:00:01.148)       0:24:34.125 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:27:11 +0000 (0:00:01.260)       0:24:35.385 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:27:13 +0000 (0:00:01.823)       0:24:37.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:27:14 +0000 (0:00:01.112)       0:24:38.321 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:27:15 +0000 (0:00:01.131)       0:24:39.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:27:16 +0000 (0:00:01.147)       0:24:40.599 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:27:18 +0000 (0:00:01.135)       0:24:41.735 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:27:19 +0000 (0:00:01.173)       0:24:42.909 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:27:20 +0000 (0:00:01.175)       0:24:44.084 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:27:23 +0000 (0:00:02.611)       0:24:46.696 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:27:24 +0000 (0:00:01.159)       0:24:47.855 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:27:25 +0000 (0:00:01.298)       0:24:49.154 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:27:26 +0000 (0:00:01.141)       0:24:50.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:27:27 +0000 (0:00:01.144)       0:24:51.440 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:27:28 +0000 (0:00:01.127)       0:24:52.568 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:27:30 +0000 (0:00:01.194)       0:24:53.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:27:31 +0000 (0:00:01.142)       0:24:54.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:27:32 +0000 (0:00:01.173)       0:24:56.079 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:27:33 +0000 (0:00:01.149)       0:24:57.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:27:34 +0000 (0:00:01.161)       0:24:58.390 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:27:35 +0000 (0:00:01.158)       0:24:59.548 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:27:37 +0000 (0:00:01.118)       0:25:00.667 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:27:44 +0000 (0:00:07.036)       0:25:07.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:27:45 +0000 (0:00:01.134)       0:25:08.837 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:27:46 +0000 (0:00:01.117)       0:25:09.955 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:27:47 +0000 (0:00:01.153)       0:25:11.109 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:27:48 +0000 (0:00:01.133)       0:25:12.243 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:27:49 +0000 (0:00:01.139)       0:25:13.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:27:50 +0000 (0:00:01.108)       0:25:14.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:27:51 +0000 (0:00:01.102)       0:25:15.593 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:27:53 +0000 (0:00:01.116)       0:25:16.710 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:27:54 +0000 (0:00:01.112)       0:25:17.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:27:55 +0000 (0:00:01.125)       0:25:18.949 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:27:56 +0000 (0:00:01.133)       0:25:20.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:27:57 +0000 (0:00:01.154)       0:25:21.237 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:27:58 +0000 (0:00:01.192)       0:25:22.429 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:27:59 +0000 (0:00:01.157)       0:25:23.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:28:01 +0000 (0:00:01.294)       0:25:24.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:28:02 +0000 (0:00:01.262)       0:25:26.143 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:28:03 +0000 (0:00:01.094)       0:25:27.238 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:28:04 +0000 (0:00:01.118)       0:25:28.357 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:28:05 +0000 (0:00:01.147)       0:25:29.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:28:06 +0000 (0:00:00.973)       0:25:30.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:28:07 +0000 (0:00:01.029)       0:25:31.507 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:28:09 +0000 (0:00:01.337)       0:25:32.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:28:10 +0000 (0:00:01.380)       0:25:34.225 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:28:12 +0000 (0:00:01.607)       0:25:35.833 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:28:13 +0000 (0:00:01.149)       0:25:36.982 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:28:14 +0000 (0:00:01.442)       0:25:38.424 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:28:16 +0000 (0:00:01.846)       0:25:40.271 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:28:18 +0000 (0:00:01.689)       0:25:41.960 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Thursday 26 March 2026  05:28:19 +0000 (0:00:01.473)       0:25:43.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Thursday 26 March 2026  05:28:21 +0000 (0:00:01.533)       0:25:44.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Thursday 26 March 2026  05:28:22 +0000 (0:00:01.134)       0:25:46.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Thursday 26 March 2026  05:28:30 +0000 (0:00:07.644)       0:25:53.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:28:31 +0000 (0:00:01.177)       0:25:54.923 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:28:34 +0000 (0:00:03.200)       0:25:58.124 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Thursday 26 March 2026  05:28:36 +0000 (0:00:02.001)       0:26:00.125 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Thursday 26 March 2026  05:28:38 +0000 (0:00:01.570)       0:26:01.696 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:28:39 +0000 (0:00:01.205)       0:26:02.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:28:40 +0000 (0:00:01.138)       0:26:04.040 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:28:41 +0000 (0:00:01.469)       0:26:05.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:28:43 +0000 (0:00:01.168)       0:26:06.679 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:28:44 +0000 (0:00:01.120)       0:26:07.800 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:28:45 +0000 (0:00:01.482)       0:26:09.283 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Thursday 26 March 2026  05:28:47 +0000 (0:00:02.048)       0:26:11.331 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Thursday 26 March 2026  05:28:50 +0000 (0:00:02.472)       0:26:13.803 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Thursday 26 March 2026  05:28:52 +0000 (0:00:02.585)       0:26:16.389 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:28:56 +0000 (0:00:03.817)       0:26:20.207 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes] **************************************************",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:28:57 +0000 (0:00:01.285)       0:26:21.492 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Mask ceph mgr systemd unit] **********************************************",
                                "Thursday 26 March 2026  05:29:00 +0000 (0:00:02.406)       0:26:23.899 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:29:02 +0000 (0:00:02.020)       0:26:25.919 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:29:03 +0000 (0:00:01.115)       0:26:27.035 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:29:04 +0000 (0:00:01.507)       0:26:28.542 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:29:06 +0000 (0:00:01.141)       0:26:29.683 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:29:07 +0000 (0:00:01.498)       0:26:31.182 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:29:08 +0000 (0:00:01.193)       0:26:32.376 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:29:09 +0000 (0:00:01.131)       0:26:33.508 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:29:10 +0000 (0:00:01.121)       0:26:34.630 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:29:12 +0000 (0:00:01.212)       0:26:35.842 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:29:13 +0000 (0:00:01.130)       0:26:36.973 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:29:15 +0000 (0:00:01.718)       0:26:38.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:29:16 +0000 (0:00:01.295)       0:26:39.988 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:29:19 +0000 (0:00:02.878)       0:26:42.867 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:29:20 +0000 (0:00:01.387)       0:26:44.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:29:22 +0000 (0:00:01.674)       0:26:45.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:29:23 +0000 (0:00:01.170)       0:26:47.099 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:29:16.907211', 'end': '2026-03-26 05:29:16.951545', 'delta': '0:00:00.044334', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:29:17.432469', 'end': '2026-03-26 05:29:17.476218', 'delta': '0:00:00.043749', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:29:17.941866', 'end': '2026-03-26 05:29:17.979790', 'delta': '0:00:00.037924', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:29:24 +0000 (0:00:01.215)       0:26:48.315 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:29:25 +0000 (0:00:01.296)       0:26:49.611 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:29:27 +0000 (0:00:01.226)       0:26:50.838 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:29:28 +0000 (0:00:01.159)       0:26:51.997 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:29:30 +0000 (0:00:02.059)       0:26:54.057 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:29:31 +0000 (0:00:01.115)       0:26:55.173 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:29:32 +0000 (0:00:01.191)       0:26:56.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:29:33 +0000 (0:00:01.258)       0:26:57.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:29:35 +0000 (0:00:01.136)       0:26:58.759 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:29:36 +0000 (0:00:01.140)       0:26:59.900 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:29:37 +0000 (0:00:01.161)       0:27:01.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:29:38 +0000 (0:00:01.142)       0:27:02.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:29:39 +0000 (0:00:01.142)       0:27:03.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:29:40 +0000 (0:00:01.148)       0:27:04.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:29:41 +0000 (0:00:01.160)       0:27:05.657 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:29:43 +0000 (0:00:01.294)       0:27:06.951 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-16-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2e41bcf9', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1', 'scsi-SQEMU_QEMU_HARDDISK_2e41bcf9-ad92-42bb-b49e-289ca95def9f-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:29:44 +0000 (0:00:01.259)       0:27:08.210 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:29:46 +0000 (0:00:01.563)       0:27:09.774 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:29:47 +0000 (0:00:01.115)       0:27:10.889 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:29:48 +0000 (0:00:01.501)       0:27:12.390 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:29:49 +0000 (0:00:01.148)       0:27:13.539 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:29:51 +0000 (0:00:01.275)       0:27:14.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:29:52 +0000 (0:00:01.166)       0:27:15.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:29:53 +0000 (0:00:01.664)       0:27:17.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:29:55 +0000 (0:00:01.160)       0:27:18.807 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:29:56 +0000 (0:00:01.127)       0:27:19.935 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:29:58 +0000 (0:00:02.117)       0:27:22.052 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:30:00 +0000 (0:00:02.294)       0:27:24.347 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:30:01 +0000 (0:00:01.124)       0:27:25.471 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:30:03 +0000 (0:00:01.317)       0:27:26.789 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:30:04 +0000 (0:00:01.521)       0:27:28.310 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:30:05 +0000 (0:00:01.166)       0:27:29.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:30:06 +0000 (0:00:01.083)       0:27:30.561 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:30:07 +0000 (0:00:01.080)       0:27:31.641 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:30:09 +0000 (0:00:01.550)       0:27:33.192 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:30:10 +0000 (0:00:01.089)       0:27:34.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:30:11 +0000 (0:00:01.108)       0:27:35.390 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:30:13 +0000 (0:00:01.582)       0:27:36.972 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:30:14 +0000 (0:00:01.528)       0:27:38.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:30:15 +0000 (0:00:00.756)       0:27:39.257 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:30:16 +0000 (0:00:00.834)       0:27:40.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:30:17 +0000 (0:00:00.774)       0:27:40.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:30:17 +0000 (0:00:00.756)       0:27:41.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:30:18 +0000 (0:00:00.810)       0:27:42.432 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:30:19 +0000 (0:00:00.836)       0:27:43.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:30:20 +0000 (0:00:00.763)       0:27:44.032 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:30:21 +0000 (0:00:00.823)       0:27:44.856 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:30:21 +0000 (0:00:00.788)       0:27:45.645 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:30:22 +0000 (0:00:00.828)       0:27:46.474 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:30:23 +0000 (0:00:00.808)       0:27:47.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:30:24 +0000 (0:00:00.780)       0:27:48.063 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:30:25 +0000 (0:00:00.780)       0:27:48.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:30:25 +0000 (0:00:00.789)       0:27:49.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:30:26 +0000 (0:00:00.777)       0:27:50.410 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:30:27 +0000 (0:00:00.784)       0:27:51.195 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:30:28 +0000 (0:00:00.825)       0:27:52.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:30:29 +0000 (0:00:00.765)       0:27:52.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:30:29 +0000 (0:00:00.747)       0:27:53.533 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:30:30 +0000 (0:00:00.772)       0:27:54.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:30:31 +0000 (0:00:00.765)       0:27:55.072 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:30:32 +0000 (0:00:00.815)       0:27:55.887 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:30:33 +0000 (0:00:01.567)       0:27:57.455 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:30:35 +0000 (0:00:01.940)       0:27:59.395 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:30:36 +0000 (0:00:01.097)       0:28:00.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:30:38 +0000 (0:00:01.191)       0:28:01.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:30:39 +0000 (0:00:01.120)       0:28:02.806 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:30:40 +0000 (0:00:01.829)       0:28:04.636 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:30:42 +0000 (0:00:01.533)       0:28:06.169 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:30:43 +0000 (0:00:01.186)       0:28:07.356 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:30:44 +0000 (0:00:00.764)       0:28:08.120 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:30:45 +0000 (0:00:00.795)       0:28:08.916 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:30:46 +0000 (0:00:01.216)       0:28:10.133 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:30:48 +0000 (0:00:01.843)       0:28:11.977 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:30:49 +0000 (0:00:01.130)       0:28:13.107 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:30:50 +0000 (0:00:01.157)       0:28:14.265 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:30:51 +0000 (0:00:01.189)       0:28:15.455 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:30:52 +0000 (0:00:01.150)       0:28:16.606 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:30:54 +0000 (0:00:01.164)       0:28:17.770 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:30:54 +0000 (0:00:00.777)       0:28:18.548 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:30:57 +0000 (0:00:02.187)       0:28:20.735 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:30:57 +0000 (0:00:00.812)       0:28:21.548 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:30:59 +0000 (0:00:01.166)       0:28:22.715 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:31:00 +0000 (0:00:01.135)       0:28:23.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:31:01 +0000 (0:00:01.143)       0:28:24.994 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:31:02 +0000 (0:00:01.145)       0:28:26.140 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:31:03 +0000 (0:00:01.120)       0:28:27.261 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:31:04 +0000 (0:00:01.113)       0:28:28.374 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:31:05 +0000 (0:00:01.153)       0:28:29.528 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:31:06 +0000 (0:00:01.126)       0:28:30.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:31:08 +0000 (0:00:01.164)       0:28:31.819 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:31:09 +0000 (0:00:00.941)       0:28:32.761 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:31:10 +0000 (0:00:01.112)       0:28:33.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:31:16 +0000 (0:00:06.138)       0:28:40.013 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:31:17 +0000 (0:00:00.765)       0:28:40.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:31:17 +0000 (0:00:00.768)       0:28:41.547 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:31:18 +0000 (0:00:00.801)       0:28:42.348 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:31:19 +0000 (0:00:00.784)       0:28:43.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:31:20 +0000 (0:00:00.771)       0:28:43.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:31:21 +0000 (0:00:00.787)       0:28:44.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:31:21 +0000 (0:00:00.750)       0:28:45.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:31:22 +0000 (0:00:00.821)       0:28:46.265 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:31:23 +0000 (0:00:00.844)       0:28:47.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:31:24 +0000 (0:00:00.797)       0:28:47.907 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:31:25 +0000 (0:00:00.772)       0:28:48.680 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:31:25 +0000 (0:00:00.766)       0:28:49.447 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:31:26 +0000 (0:00:00.900)       0:28:50.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:31:27 +0000 (0:00:00.764)       0:28:51.112 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:31:28 +0000 (0:00:00.869)       0:28:51.981 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:31:29 +0000 (0:00:00.761)       0:28:52.743 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:31:29 +0000 (0:00:00.772)       0:28:53.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:31:30 +0000 (0:00:00.821)       0:28:54.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:31:31 +0000 (0:00:00.844)       0:28:55.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:31:32 +0000 (0:00:00.826)       0:28:56.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:31:33 +0000 (0:00:00.795)       0:28:56.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:31:34 +0000 (0:00:01.387)       0:28:58.192 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:31:36 +0000 (0:00:01.467)       0:28:59.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:31:37 +0000 (0:00:01.071)       0:29:00.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:31:37 +0000 (0:00:00.773)       0:29:01.504 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:31:38 +0000 (0:00:00.894)       0:29:02.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:31:40 +0000 (0:00:01.430)       0:29:03.829 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:31:41 +0000 (0:00:01.332)       0:29:05.162 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Thursday 26 March 2026  05:31:42 +0000 (0:00:01.152)       0:29:06.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Thursday 26 March 2026  05:31:44 +0000 (0:00:01.495)       0:29:07.809 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Thursday 26 March 2026  05:31:45 +0000 (0:00:01.158)       0:29:08.968 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Thursday 26 March 2026  05:31:53 +0000 (0:00:07.848)       0:29:16.817 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:31:54 +0000 (0:00:01.260)       0:29:18.078 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:31:57 +0000 (0:00:03.250)       0:29:21.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Thursday 26 March 2026  05:31:59 +0000 (0:00:02.080)       0:29:23.409 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Thursday 26 March 2026  05:32:01 +0000 (0:00:01.496)       0:29:24.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:32:02 +0000 (0:00:00.755)       0:29:25.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:32:02 +0000 (0:00:00.758)       0:29:26.418 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:32:03 +0000 (0:00:01.114)       0:29:27.533 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:32:04 +0000 (0:00:01.122)       0:29:28.655 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:32:06 +0000 (0:00:01.116)       0:29:29.772 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:32:07 +0000 (0:00:01.096)       0:29:30.868 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Thursday 26 March 2026  05:32:09 +0000 (0:00:02.028)       0:29:32.897 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Thursday 26 March 2026  05:32:11 +0000 (0:00:01.977)       0:29:34.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Thursday 26 March 2026  05:32:13 +0000 (0:00:02.354)       0:29:37.229 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:32:16 +0000 (0:00:03.301)       0:29:40.531 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mgr nodes] **************************************************",
                                "",
                                "TASK [Stop ceph mgr] ***********************************************************",
                                "Thursday 26 March 2026  05:32:17 +0000 (0:00:01.014)       0:29:41.546 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Mask ceph mgr systemd unit] **********************************************",
                                "Thursday 26 March 2026  05:32:20 +0000 (0:00:02.570)       0:29:44.116 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:32:22 +0000 (0:00:02.265)       0:29:46.381 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:32:23 +0000 (0:00:01.153)       0:29:47.534 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:32:25 +0000 (0:00:01.508)       0:29:49.042 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:32:26 +0000 (0:00:01.102)       0:29:50.145 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:32:27 +0000 (0:00:01.484)       0:29:51.630 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:32:29 +0000 (0:00:01.167)       0:29:52.798 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:32:30 +0000 (0:00:01.131)       0:29:53.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:32:31 +0000 (0:00:01.195)       0:29:55.125 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:32:32 +0000 (0:00:01.131)       0:29:56.256 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:32:33 +0000 (0:00:01.164)       0:29:57.421 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:32:35 +0000 (0:00:01.704)       0:29:59.126 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:32:36 +0000 (0:00:01.260)       0:30:00.386 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:32:39 +0000 (0:00:02.884)       0:30:03.271 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:32:41 +0000 (0:00:01.399)       0:30:04.671 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:32:43 +0000 (0:00:01.993)       0:30:06.665 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:32:44 +0000 (0:00:01.221)       0:30:07.887 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:32:37.251347', 'end': '2026-03-26 05:32:37.306087', 'delta': '0:00:00.054740', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:32:37.856920', 'end': '2026-03-26 05:32:37.897986', 'delta': '0:00:00.041066', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:32:38.441829', 'end': '2026-03-26 05:32:38.485013', 'delta': '0:00:00.043184', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:32:45 +0000 (0:00:01.227)       0:30:09.114 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:32:46 +0000 (0:00:01.275)       0:30:10.389 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:32:48 +0000 (0:00:01.670)       0:30:12.060 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:32:49 +0000 (0:00:01.143)       0:30:13.204 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:32:51 +0000 (0:00:01.949)       0:30:15.154 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:32:52 +0000 (0:00:01.128)       0:30:16.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:32:53 +0000 (0:00:01.152)       0:30:17.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:32:55 +0000 (0:00:01.237)       0:30:18.673 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:32:56 +0000 (0:00:01.163)       0:30:19.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:32:57 +0000 (0:00:01.154)       0:30:20.991 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:32:58 +0000 (0:00:01.156)       0:30:22.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:32:59 +0000 (0:00:01.121)       0:30:23.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:33:00 +0000 (0:00:01.146)       0:30:24.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:33:01 +0000 (0:00:01.159)       0:30:25.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:33:03 +0000 (0:00:01.175)       0:30:26.752 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:33:04 +0000 (0:00:01.292)       0:30:28.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-10-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7634648a', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1', 'scsi-SQEMU_QEMU_HARDDISK_7634648a-b5a4-45bc-ac0b-8484a2642b22-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:33:05 +0000 (0:00:01.278)       0:30:29.323 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:33:07 +0000 (0:00:01.481)       0:30:30.804 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:33:08 +0000 (0:00:01.158)       0:30:31.962 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:33:09 +0000 (0:00:01.491)       0:30:33.454 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:33:10 +0000 (0:00:01.157)       0:30:34.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:33:12 +0000 (0:00:01.334)       0:30:35.947 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:33:13 +0000 (0:00:01.185)       0:30:37.132 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:33:15 +0000 (0:00:01.723)       0:30:38.856 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:33:16 +0000 (0:00:01.165)       0:30:40.022 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:33:17 +0000 (0:00:01.127)       0:30:41.150 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:33:19 +0000 (0:00:02.171)       0:30:43.322 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:33:21 +0000 (0:00:02.279)       0:30:45.601 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:33:23 +0000 (0:00:01.268)       0:30:46.870 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:33:24 +0000 (0:00:01.121)       0:30:47.991 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:33:25 +0000 (0:00:01.538)       0:30:49.530 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:33:27 +0000 (0:00:01.136)       0:30:50.667 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:33:28 +0000 (0:00:01.098)       0:30:51.765 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:33:29 +0000 (0:00:01.166)       0:30:52.932 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:33:30 +0000 (0:00:01.577)       0:30:54.509 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:33:32 +0000 (0:00:01.202)       0:30:55.712 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:33:33 +0000 (0:00:01.162)       0:30:56.875 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:33:34 +0000 (0:00:01.605)       0:30:58.480 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:33:36 +0000 (0:00:01.614)       0:31:00.095 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:33:37 +0000 (0:00:00.980)       0:31:01.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:33:38 +0000 (0:00:00.858)       0:31:01.934 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:33:39 +0000 (0:00:00.773)       0:31:02.707 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:33:39 +0000 (0:00:00.810)       0:31:03.518 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:33:40 +0000 (0:00:00.804)       0:31:04.322 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:33:41 +0000 (0:00:00.860)       0:31:05.183 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:33:42 +0000 (0:00:00.803)       0:31:05.986 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:33:43 +0000 (0:00:00.824)       0:31:06.811 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:33:44 +0000 (0:00:00.851)       0:31:07.662 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:33:44 +0000 (0:00:00.886)       0:31:08.549 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:33:45 +0000 (0:00:00.794)       0:31:09.343 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:33:46 +0000 (0:00:00.869)       0:31:10.213 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:33:47 +0000 (0:00:01.025)       0:31:11.239 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:33:48 +0000 (0:00:00.798)       0:31:12.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:33:49 +0000 (0:00:00.777)       0:31:12.816 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:33:49 +0000 (0:00:00.792)       0:31:13.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:33:50 +0000 (0:00:00.835)       0:31:14.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:33:51 +0000 (0:00:00.774)       0:31:15.218 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:33:52 +0000 (0:00:00.815)       0:31:16.034 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:33:53 +0000 (0:00:00.779)       0:31:16.813 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:33:53 +0000 (0:00:00.810)       0:31:17.623 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:33:54 +0000 (0:00:00.791)       0:31:18.415 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:33:56 +0000 (0:00:01.710)       0:31:20.125 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:33:58 +0000 (0:00:02.181)       0:31:22.307 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:34:00 +0000 (0:00:01.450)       0:31:23.758 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:34:01 +0000 (0:00:01.168)       0:31:24.926 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:34:02 +0000 (0:00:01.150)       0:31:26.077 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:34:04 +0000 (0:00:01.875)       0:31:27.952 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:34:05 +0000 (0:00:01.473)       0:31:29.426 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:34:06 +0000 (0:00:01.203)       0:31:30.629 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:34:07 +0000 (0:00:00.788)       0:31:31.417 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:34:08 +0000 (0:00:00.772)       0:31:32.190 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:34:09 +0000 (0:00:01.122)       0:31:33.313 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:34:11 +0000 (0:00:01.927)       0:31:35.241 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:34:12 +0000 (0:00:01.240)       0:31:36.481 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:34:14 +0000 (0:00:01.219)       0:31:37.700 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:34:15 +0000 (0:00:01.326)       0:31:39.027 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:34:16 +0000 (0:00:01.216)       0:31:40.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:34:17 +0000 (0:00:01.305)       0:31:41.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:34:18 +0000 (0:00:00.847)       0:31:42.398 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:34:21 +0000 (0:00:02.272)       0:31:44.670 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:34:21 +0000 (0:00:00.828)       0:31:45.498 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:34:23 +0000 (0:00:01.177)       0:31:46.676 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:34:24 +0000 (0:00:01.196)       0:31:47.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:34:25 +0000 (0:00:01.171)       0:31:49.044 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:34:26 +0000 (0:00:01.132)       0:31:50.177 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:34:27 +0000 (0:00:01.159)       0:31:51.336 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:34:28 +0000 (0:00:01.211)       0:31:52.548 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:34:30 +0000 (0:00:01.324)       0:31:53.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:34:31 +0000 (0:00:01.190)       0:31:55.063 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:34:32 +0000 (0:00:01.148)       0:31:56.212 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:34:33 +0000 (0:00:00.819)       0:31:57.032 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:34:34 +0000 (0:00:01.127)       0:31:58.159 ******** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:34:40 +0000 (0:00:06.437)       0:32:04.596 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:34:41 +0000 (0:00:00.780)       0:32:05.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:34:42 +0000 (0:00:00.839)       0:32:06.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:34:43 +0000 (0:00:00.808)       0:32:07.025 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:34:44 +0000 (0:00:00.793)       0:32:07.819 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:34:44 +0000 (0:00:00.793)       0:32:08.613 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:34:45 +0000 (0:00:00.823)       0:32:09.436 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:34:46 +0000 (0:00:00.753)       0:32:10.190 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:34:47 +0000 (0:00:00.769)       0:32:10.959 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:34:48 +0000 (0:00:00.786)       0:32:11.745 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:34:48 +0000 (0:00:00.787)       0:32:12.533 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:34:49 +0000 (0:00:00.806)       0:32:13.340 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:34:50 +0000 (0:00:00.770)       0:32:14.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:34:51 +0000 (0:00:00.902)       0:32:15.013 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:34:52 +0000 (0:00:00.785)       0:32:15.798 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:34:53 +0000 (0:00:00.930)       0:32:16.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:34:53 +0000 (0:00:00.765)       0:32:17.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:34:54 +0000 (0:00:00.787)       0:32:18.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:34:55 +0000 (0:00:00.814)       0:32:19.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:34:56 +0000 (0:00:00.816)       0:32:19.914 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:34:57 +0000 (0:00:00.802)       0:32:20.716 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:34:57 +0000 (0:00:00.778)       0:32:21.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:34:58 +0000 (0:00:01.097)       0:32:22.592 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:34:59 +0000 (0:00:01.047)       0:32:23.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:35:01 +0000 (0:00:01.127)       0:32:24.767 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:35:01 +0000 (0:00:00.815)       0:32:25.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:35:02 +0000 (0:00:00.910)       0:32:26.493 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:35:04 +0000 (0:00:01.445)       0:32:27.938 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:35:05 +0000 (0:00:01.653)       0:32:29.592 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Thursday 26 March 2026  05:35:07 +0000 (0:00:01.119)       0:32:30.712 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Thursday 26 March 2026  05:35:08 +0000 (0:00:01.472)       0:32:32.185 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Thursday 26 March 2026  05:35:09 +0000 (0:00:01.102)       0:32:33.287 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Thursday 26 March 2026  05:35:16 +0000 (0:00:07.050)       0:32:40.338 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:35:17 +0000 (0:00:01.191)       0:32:41.530 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:35:21 +0000 (0:00:03.340)       0:32:44.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Thursday 26 March 2026  05:35:23 +0000 (0:00:02.059)       0:32:46.929 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Thursday 26 March 2026  05:35:24 +0000 (0:00:01.520)       0:32:48.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Thursday 26 March 2026  05:35:25 +0000 (0:00:00.803)       0:32:49.253 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Thursday 26 March 2026  05:35:26 +0000 (0:00:00.794)       0:32:50.048 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:35:27 +0000 (0:00:01.099)       0:32:51.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:35:28 +0000 (0:00:01.222)       0:32:52.371 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:35:29 +0000 (0:00:01.136)       0:32:53.507 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:35:31 +0000 (0:00:01.243)       0:32:54.751 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Thursday 26 March 2026  05:35:33 +0000 (0:00:02.052)       0:32:56.803 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Thursday 26 March 2026  05:35:35 +0000 (0:00:01.978)       0:32:58.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Thursday 26 March 2026  05:35:37 +0000 (0:00:02.506)       0:33:01.289 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Thursday 26 March 2026  05:35:41 +0000 (0:00:03.625)       0:33:04.914 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Thursday 26 March 2026  05:35:42 +0000 (0:00:01.571)       0:33:06.486 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Thursday 26 March 2026  05:35:45 +0000 (0:00:02.483)       0:33:08.970 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Thursday 26 March 2026  05:35:47 +0000 (0:00:02.344)       0:33:11.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Thursday 26 March 2026  05:35:48 +0000 (0:00:01.318)       0:33:12.632 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Thursday 26 March 2026  05:35:50 +0000 (0:00:01.156)       0:33:13.789 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=dashboard) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=prometheus) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Thursday 26 March 2026  05:35:51 +0000 (0:00:01.713)       0:33:15.502 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=dashboard) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=prometheus) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Set osd flags] ***********************************************************",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:35:53 +0000 (0:00:01.983)       0:33:17.486 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:35:55 +0000 (0:00:01.787)       0:33:19.274 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get pool list] ***********************************************************",
                                "Thursday 26 March 2026  05:35:57 +0000 (0:00:01.730)       0:33:21.004 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Get balancer module status] **********************************************",
                                "Thursday 26 March 2026  05:36:00 +0000 (0:00:02.915)       0:33:23.920 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Set_fact pools_pgautoscaler_mode] ****************************************",
                                "Thursday 26 March 2026  05:36:03 +0000 (0:00:02.829)       0:33:26.749 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 1, 'pool_name': '.mgr', 'create_time': '2026-03-26T02:57:09.996917+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 2, 'min_size': 1, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_acting': 6.059999942779541, 'score_stable': 6.059999942779541, 'optimal_score': 0.33000001311302185, 'raw_score_acting': 2, 'raw_score_stable': 2, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 2, 'pool_name': 'cephfs_data', 'create_time': '2026-03-26T02:58:21.868811+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '32', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '30', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'cephfs': {'data': 'cephfs'}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 3, 'pool_name': 'cephfs_metadata', 'create_time': '2026-03-26T02:58:25.656609+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 16, 'pg_placement_num': 16, 'pg_placement_num_target': 16, 'pg_num_target': 16, 'pg_num_pending': 16, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '64', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '30', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_autoscale_bias': 4, 'pg_num_min': 16, 'recovery_priority': 5}, 'application_metadata': {'cephfs': {'metadata': 'cephfs'}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 4, 'pool_name': 'default.rgw.buckets.data', 'create_time': '2026-03-26T02:59:25.200032+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '69', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 5, 'pool_name': 'default.rgw.buckets.index', 'create_time': '2026-03-26T02:59:31.330922+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '71', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 6, 'pool_name': 'default.rgw.control', 'create_time': '2026-03-26T02:59:37.488224+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '71', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.8799999952316284, 'score_stable': 1.8799999952316284, 'optimal_score': 1, 'raw_score_acting': 1.8799999952316284, 'raw_score_stable': 1.8799999952316284, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 7, 'pool_name': 'default.rgw.log', 'create_time': '2026-03-26T02:59:43.721154+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '194', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '73', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 8, 'pool_name': 'default.rgw.meta', 'create_time': '2026-03-26T02:59:50.241088+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '73', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 9, 'pool_name': '.rgw.root', 'create_time': '2026-03-26T03:00:02.559162+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 2, 'min_size': 1, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'on', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '77', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '75', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rgw': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 10, 'pool_name': 'backups', 'create_time': '2026-03-26T03:00:50.660723+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '102', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 102, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 2.059999942779541, 'score_stable': 2.059999942779541, 'optimal_score': 1, 'raw_score_acting': 2.059999942779541, 'raw_score_stable': 2.059999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 11, 'pool_name': 'volumes', 'create_time': '2026-03-26T03:01:00.012041+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '109', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 109, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 12, 'pool_name': 'images', 'create_time': '2026-03-26T03:01:10.076355+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '204', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 6, 'snap_epoch': 204, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.309999942779541, 'score_stable': 1.309999942779541, 'optimal_score': 1, 'raw_score_acting': 1.309999942779541, 'raw_score_stable': 1.309999942779541, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 13, 'pool_name': 'metrics', 'create_time': '2026-03-26T03:01:18.926005+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '125', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 125, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.690000057220459, 'score_stable': 1.690000057220459, 'optimal_score': 1, 'raw_score_acting': 1.690000057220459, 'raw_score_stable': 1.690000057220459, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'pool_id': 14, 'pool_name': 'vms', 'create_time': '2026-03-26T03:01:27.854751+0000', 'flags': 8193, 'flags_names': 'hashpspool,selfmanaged_snaps', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 32, 'pg_placement_num': 32, 'pg_placement_num_target': 32, 'pg_num_target': 32, 'pg_num_pending': 32, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': \"0'0\", 'target_version': \"0'0\"}, 'last_change': '132', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 3, 'snap_epoch': 132, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {}, 'application_metadata': {'rbd': {}}, 'read_balance': {'score_acting': 1.5, 'score_stable': 1.5, 'optimal_score': 1, 'raw_score_acting': 1.5, 'raw_score_stable': 1.5, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}})\u001b[0m",
                                "",
                                "TASK [Disable balancer] ********************************************************",
                                "Thursday 26 March 2026  05:36:05 +0000 (0:00:02.826)       0:33:29.575 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Disable pg autoscale on pools] *******************************************",
                                "Thursday 26 March 2026  05:36:08 +0000 (0:00:03.027)       0:33:32.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.mgr', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_data', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_metadata', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.data', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.index', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.control', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.log', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.meta', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.rgw.root', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'backups', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'volumes', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'images', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'metrics', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vms', 'mode': 'off'}) \u001b[0m",
                                "",
                                "TASK [Set osd flags] ***********************************************************",
                                "Thursday 26 March 2026  05:37:23 +0000 (0:01:14.542)       0:34:47.145 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=noout)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=nodeep-scrub)\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph osds cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:37:29 +0000 (0:00:06.111)       0:34:53.257 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:37:30 +0000 (0:00:01.142)       0:34:54.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:37:32 +0000 (0:00:01.510)       0:34:55.910 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:37:33 +0000 (0:00:01.132)       0:34:57.043 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:37:34 +0000 (0:00:01.490)       0:34:58.534 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:37:36 +0000 (0:00:01.169)       0:34:59.703 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:37:37 +0000 (0:00:01.138)       0:35:00.842 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:37:38 +0000 (0:00:01.145)       0:35:01.987 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:37:39 +0000 (0:00:01.166)       0:35:03.153 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:37:40 +0000 (0:00:01.136)       0:35:04.290 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:37:42 +0000 (0:00:01.695)       0:35:05.986 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:37:43 +0000 (0:00:01.308)       0:35:07.294 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:37:46 +0000 (0:00:03.319)       0:35:10.613 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:37:48 +0000 (0:00:01.439)       0:35:12.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:37:50 +0000 (0:00:01.989)       0:35:14.043 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:37:51 +0000 (0:00:01.182)       0:35:15.226 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:37:44.202827', 'end': '2026-03-26 05:37:44.263426', 'delta': '0:00:00.060599', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:37:44.808591', 'end': '2026-03-26 05:37:44.855653', 'delta': '0:00:00.047062', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:37:45.729323', 'end': '2026-03-26 05:37:45.782176', 'delta': '0:00:00.052853', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:37:52 +0000 (0:00:01.196)       0:35:16.423 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:37:54 +0000 (0:00:01.672)       0:35:18.095 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:37:55 +0000 (0:00:01.266)       0:35:19.362 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:37:56 +0000 (0:00:01.164)       0:35:20.527 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:37:58 +0000 (0:00:01.979)       0:35:22.507 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:38:00 +0000 (0:00:01.155)       0:35:23.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:38:01 +0000 (0:00:01.146)       0:35:24.809 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:38:02 +0000 (0:00:01.225)       0:35:26.034 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:38:03 +0000 (0:00:01.121)       0:35:27.155 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:38:04 +0000 (0:00:01.114)       0:35:28.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:38:05 +0000 (0:00:01.199)       0:35:29.469 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:38:06 +0000 (0:00:01.152)       0:35:30.622 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:38:08 +0000 (0:00:01.200)       0:35:31.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:38:09 +0000 (0:00:01.118)       0:35:32.941 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:38:10 +0000 (0:00:01.135)       0:35:34.076 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:38:12 +0000 (0:00:01.740)       0:35:35.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:38:13 +0000 (0:00:01.360)       0:35:37.178 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:38:15 +0000 (0:00:01.523)       0:35:38.701 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:38:16 +0000 (0:00:01.114)       0:35:39.816 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:38:17 +0000 (0:00:01.470)       0:35:41.286 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:38:18 +0000 (0:00:01.100)       0:35:42.386 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:38:19 +0000 (0:00:01.257)       0:35:43.644 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:38:21 +0000 (0:00:01.126)       0:35:44.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:38:23 +0000 (0:00:02.083)       0:35:46.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:38:24 +0000 (0:00:01.240)       0:35:48.095 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:38:25 +0000 (0:00:01.138)       0:35:49.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:38:26 +0000 (0:00:01.171)       0:35:50.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:38:27 +0000 (0:00:01.135)       0:35:51.539 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:38:29 +0000 (0:00:01.169)       0:35:52.709 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:38:30 +0000 (0:00:01.234)       0:35:53.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:38:31 +0000 (0:00:01.411)       0:35:55.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:38:33 +0000 (0:00:01.424)       0:35:56.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:38:34 +0000 (0:00:01.373)       0:35:58.153 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:38:35 +0000 (0:00:01.143)       0:35:59.296 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:38:36 +0000 (0:00:01.352)       0:36:00.649 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:38:39 +0000 (0:00:02.178)       0:36:02.827 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get osd numbers - non container] *****************************************",
                                "Thursday 26 March 2026  05:38:41 +0000 (0:00:02.664)       0:36:05.492 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set num_osds] ************************************************************",
                                "Thursday 26 March 2026  05:38:43 +0000 (0:00:01.525)       0:36:07.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set_fact container_exec_cmd_osd] *****************************************",
                                "Thursday 26 March 2026  05:38:44 +0000 (0:00:01.180)       0:36:08.198 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Stop ceph osd] ***********************************************************",
                                "Thursday 26 March 2026  05:38:46 +0000 (0:00:01.655)       0:36:09.853 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:38:50 +0000 (0:00:04.297)       0:36:14.150 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:38:51 +0000 (0:00:01.179)       0:36:15.330 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:38:52 +0000 (0:00:01.143)       0:36:16.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:38:54 +0000 (0:00:01.211)       0:36:17.684 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:38:55 +0000 (0:00:01.518)       0:36:19.202 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:38:57 +0000 (0:00:01.483)       0:36:20.686 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:38:58 +0000 (0:00:01.503)       0:36:22.190 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:38:59 +0000 (0:00:01.219)       0:36:23.410 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:39:00 +0000 (0:00:01.172)       0:36:24.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:39:02 +0000 (0:00:01.125)       0:36:25.708 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:39:03 +0000 (0:00:01.514)       0:36:27.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:39:05 +0000 (0:00:01.505)       0:36:28.727 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:39:06 +0000 (0:00:01.160)       0:36:29.888 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:39:07 +0000 (0:00:01.140)       0:36:31.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:39:08 +0000 (0:00:01.157)       0:36:32.187 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:39:09 +0000 (0:00:01.181)       0:36:33.368 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:39:10 +0000 (0:00:01.148)       0:36:34.517 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:39:11 +0000 (0:00:01.102)       0:36:35.620 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:39:13 +0000 (0:00:01.140)       0:36:36.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:39:14 +0000 (0:00:01.137)       0:36:37.898 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:39:15 +0000 (0:00:01.200)       0:36:39.098 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:39:16 +0000 (0:00:01.166)       0:36:40.265 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:39:17 +0000 (0:00:01.212)       0:36:41.477 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:39:18 +0000 (0:00:01.155)       0:36:42.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:39:20 +0000 (0:00:01.115)       0:36:43.749 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:39:21 +0000 (0:00:01.125)       0:36:44.875 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:39:22 +0000 (0:00:01.151)       0:36:46.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:39:23 +0000 (0:00:01.174)       0:36:47.200 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:39:24 +0000 (0:00:01.102)       0:36:48.303 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:39:25 +0000 (0:00:01.124)       0:36:49.428 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:39:26 +0000 (0:00:01.159)       0:36:50.587 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:39:28 +0000 (0:00:01.134)       0:36:51.722 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:39:29 +0000 (0:00:01.177)       0:36:52.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:39:30 +0000 (0:00:01.104)       0:36:54.004 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:39:32 +0000 (0:00:01.971)       0:36:55.975 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:39:34 +0000 (0:00:02.286)       0:36:58.262 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:39:35 +0000 (0:00:01.153)       0:36:59.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:39:36 +0000 (0:00:01.125)       0:37:00.541 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:39:38 +0000 (0:00:01.155)       0:37:01.697 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:39:39 +0000 (0:00:01.805)       0:37:03.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:39:41 +0000 (0:00:01.514)       0:37:05.019 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:39:42 +0000 (0:00:01.228)       0:37:06.248 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:39:43 +0000 (0:00:01.235)       0:37:07.484 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:39:44 +0000 (0:00:01.158)       0:37:08.643 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:39:46 +0000 (0:00:01.111)       0:37:09.755 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:39:48 +0000 (0:00:02.889)       0:37:12.644 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:39:50 +0000 (0:00:01.176)       0:37:13.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:39:51 +0000 (0:00:01.179)       0:37:15.001 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:39:52 +0000 (0:00:01.146)       0:37:16.147 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:39:53 +0000 (0:00:01.167)       0:37:17.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:39:54 +0000 (0:00:01.178)       0:37:18.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:39:56 +0000 (0:00:01.183)       0:37:19.679 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:39:58 +0000 (0:00:02.581)       0:37:22.261 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:39:59 +0000 (0:00:01.108)       0:37:23.369 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:40:00 +0000 (0:00:01.098)       0:37:24.467 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:40:01 +0000 (0:00:01.142)       0:37:25.610 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:40:03 +0000 (0:00:01.203)       0:37:26.814 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:40:04 +0000 (0:00:01.138)       0:37:27.952 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:40:05 +0000 (0:00:01.184)       0:37:29.137 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:40:06 +0000 (0:00:01.131)       0:37:30.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:40:07 +0000 (0:00:01.130)       0:37:31.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:40:08 +0000 (0:00:01.231)       0:37:32.632 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:40:10 +0000 (0:00:01.102)       0:37:33.734 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:40:11 +0000 (0:00:01.175)       0:37:34.909 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:40:12 +0000 (0:00:01.146)       0:37:36.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:40:19 +0000 (0:00:06.674)       0:37:42.731 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  05:40:20 +0000 (0:00:01.506)       0:37:44.238 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  05:40:22 +0000 (0:00:01.533)       0:37:45.772 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:40:24 +0000 (0:00:02.002)       0:37:47.774 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:40:25 +0000 (0:00:01.145)       0:37:48.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:40:26 +0000 (0:00:01.112)       0:37:50.032 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:40:27 +0000 (0:00:01.206)       0:37:51.239 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:40:28 +0000 (0:00:01.179)       0:37:52.418 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:40:29 +0000 (0:00:01.159)       0:37:53.578 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:40:31 +0000 (0:00:01.197)       0:37:54.776 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:40:32 +0000 (0:00:01.126)       0:37:55.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:40:33 +0000 (0:00:01.176)       0:37:57.079 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:40:34 +0000 (0:00:01.154)       0:37:58.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:40:35 +0000 (0:00:01.136)       0:37:59.370 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:40:36 +0000 (0:00:01.263)       0:38:00.634 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:40:41 +0000 (0:00:04.522)       0:38:05.157 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:40:42 +0000 (0:00:01.245)       0:38:06.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:40:50 +0000 (0:00:08.009)       0:38:14.412 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:40:51 +0000 (0:00:01.139)       0:38:15.552 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:40:53 +0000 (0:00:01.147)       0:38:16.700 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:40:54 +0000 (0:00:01.178)       0:38:17.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:40:55 +0000 (0:00:01.214)       0:38:19.093 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:40:56 +0000 (0:00:01.137)       0:38:20.231 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:40:57 +0000 (0:00:01.307)       0:38:21.539 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:40:59 +0000 (0:00:01.448)       0:38:22.988 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:41:00 +0000 (0:00:01.419)       0:38:24.408 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:41:02 +0000 (0:00:01.516)       0:38:25.924 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:41:03 +0000 (0:00:01.181)       0:38:27.105 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:41:05 +0000 (0:00:01.916)       0:38:29.022 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Thursday 26 March 2026  05:41:07 +0000 (0:00:01.719)       0:38:30.741 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:41:08 +0000 (0:00:01.130)       0:38:31.871 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Thursday 26 March 2026  05:41:09 +0000 (0:00:01.668)       0:38:33.540 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Thursday 26 March 2026  05:41:11 +0000 (0:00:01.485)       0:38:35.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Thursday 26 March 2026  05:41:12 +0000 (0:00:01.114)       0:38:36.141 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Thursday 26 March 2026  05:41:13 +0000 (0:00:01.116)       0:38:37.258 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Thursday 26 March 2026  05:41:15 +0000 (0:00:01.434)       0:38:38.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Thursday 26 March 2026  05:41:16 +0000 (0:00:01.128)       0:38:39.820 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Thursday 26 March 2026  05:41:19 +0000 (0:00:02.994)       0:38:42.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Thursday 26 March 2026  05:41:20 +0000 (0:00:01.106)       0:38:43.922 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Thursday 26 March 2026  05:41:21 +0000 (0:00:01.566)       0:38:45.489 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:41:23 +0000 (0:00:01.826)       0:38:47.315 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:41:26 +0000 (0:00:03.222)       0:38:50.538 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Thursday 26 March 2026  05:41:28 +0000 (0:00:01.983)       0:38:52.522 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Thursday 26 March 2026  05:41:30 +0000 (0:00:01.247)       0:38:53.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Thursday 26 March 2026  05:41:31 +0000 (0:00:01.140)       0:38:54.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Thursday 26 March 2026  05:41:32 +0000 (0:00:01.191)       0:38:56.101 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Thursday 26 March 2026  05:41:33 +0000 (0:00:01.504)       0:38:57.606 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Thursday 26 March 2026  05:41:35 +0000 (0:00:01.458)       0:38:59.065 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:41:38 +0000 (0:00:03.593)       0:39:02.659 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:41:40 +0000 (0:00:01.484)       0:39:04.143 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Thursday 26 March 2026  05:41:42 +0000 (0:00:01.987)       0:39:06.130 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Thursday 26 March 2026  05:41:44 +0000 (0:00:01.964)       0:39:08.095 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:41:46 +0000 (0:00:02.264)       0:39:10.359 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:41:47 +0000 (0:00:01.213)       0:39:11.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Thursday 26 March 2026  05:41:49 +0000 (0:00:01.120)       0:39:12.694 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Thursday 26 March 2026  05:41:50 +0000 (0:00:01.870)       0:39:14.564 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Thursday 26 March 2026  05:41:53 +0000 (0:00:02.955)       0:39:17.520 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Thursday 26 March 2026  05:41:58 +0000 (0:00:04.710)       0:39:22.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Thursday 26 March 2026  05:41:59 +0000 (0:00:01.260)       0:39:23.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Thursday 26 March 2026  05:42:01 +0000 (0:00:01.234)       0:39:24.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************",
                                "Thursday 26 March 2026  05:42:02 +0000 (0:00:01.270)       0:39:25.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***",
                                "Thursday 26 March 2026  05:42:03 +0000 (0:00:01.154)       0:39:27.150 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Waiting for clean pgs...] ************************************************",
                                "Thursday 26 March 2026  05:42:04 +0000 (0:00:01.204)       0:39:28.355 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (600 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (599 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (598 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for clean pgs... (597 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:42:18 +0000 (0:00:14.164)       0:39:42.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:42:19 +0000 (0:00:01.121)       0:39:43.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:42:21 +0000 (0:00:01.130)       0:39:44.772 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:42:22 +0000 (0:00:01.156)       0:39:45.928 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:42:23 +0000 (0:00:01.144)       0:39:47.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:42:24 +0000 (0:00:01.141)       0:39:48.215 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:42:25 +0000 (0:00:01.107)       0:39:49.322 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph osds cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:42:26 +0000 (0:00:00.953)       0:39:50.276 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:42:27 +0000 (0:00:01.167)       0:39:51.444 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:42:29 +0000 (0:00:01.481)       0:39:52.925 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:42:30 +0000 (0:00:01.166)       0:39:54.092 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:42:31 +0000 (0:00:01.452)       0:39:55.545 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:42:33 +0000 (0:00:01.180)       0:39:56.726 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:42:34 +0000 (0:00:01.124)       0:39:57.850 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:42:35 +0000 (0:00:01.144)       0:39:58.995 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:42:36 +0000 (0:00:01.137)       0:40:00.132 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:42:37 +0000 (0:00:01.117)       0:40:01.250 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:42:39 +0000 (0:00:02.101)       0:40:03.352 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:42:40 +0000 (0:00:01.253)       0:40:04.605 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:42:44 +0000 (0:00:03.297)       0:40:07.903 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:42:46 +0000 (0:00:01.828)       0:40:09.732 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:42:48 +0000 (0:00:01.963)       0:40:11.695 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:42:49 +0000 (0:00:01.239)       0:40:12.935 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:42:41.492746', 'end': '2026-03-26 05:42:41.541679', 'delta': '0:00:00.048933', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:42:42.438597', 'end': '2026-03-26 05:42:42.480128', 'delta': '0:00:00.041531', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:42:43.005995', 'end': '2026-03-26 05:42:43.056235', 'delta': '0:00:00.050240', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:42:50 +0000 (0:00:01.178)       0:40:14.113 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:42:51 +0000 (0:00:01.281)       0:40:15.394 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:42:53 +0000 (0:00:01.272)       0:40:16.667 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:42:54 +0000 (0:00:01.164)       0:40:17.831 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:42:56 +0000 (0:00:01.961)       0:40:19.793 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:42:57 +0000 (0:00:01.116)       0:40:20.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:42:58 +0000 (0:00:01.143)       0:40:22.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:42:59 +0000 (0:00:01.241)       0:40:23.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:43:00 +0000 (0:00:01.131)       0:40:24.426 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:43:01 +0000 (0:00:01.156)       0:40:25.583 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:43:03 +0000 (0:00:01.167)       0:40:26.751 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:43:04 +0000 (0:00:01.113)       0:40:27.865 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:43:05 +0000 (0:00:01.324)       0:40:29.190 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:43:06 +0000 (0:00:01.118)       0:40:30.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:43:07 +0000 (0:00:01.123)       0:40:31.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:43:09 +0000 (0:00:01.435)       0:40:32.866 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:43:10 +0000 (0:00:01.431)       0:40:34.298 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:43:12 +0000 (0:00:01.510)       0:40:35.808 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:43:13 +0000 (0:00:01.121)       0:40:36.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:43:14 +0000 (0:00:01.521)       0:40:38.451 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:43:15 +0000 (0:00:01.121)       0:40:39.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:43:17 +0000 (0:00:01.231)       0:40:40.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:43:18 +0000 (0:00:01.132)       0:40:41.938 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:43:20 +0000 (0:00:02.059)       0:40:43.997 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:43:21 +0000 (0:00:01.160)       0:40:45.158 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:43:22 +0000 (0:00:01.262)       0:40:46.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:43:23 +0000 (0:00:01.159)       0:40:47.581 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:43:25 +0000 (0:00:01.127)       0:40:48.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:43:26 +0000 (0:00:01.137)       0:40:49.845 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:43:27 +0000 (0:00:01.247)       0:40:51.093 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:43:28 +0000 (0:00:01.412)       0:40:52.506 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:43:30 +0000 (0:00:01.419)       0:40:53.925 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:43:31 +0000 (0:00:01.426)       0:40:55.352 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:43:32 +0000 (0:00:01.248)       0:40:56.600 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:43:34 +0000 (0:00:01.457)       0:40:58.057 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:43:36 +0000 (0:00:02.195)       0:41:00.253 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get osd numbers - non container] *****************************************",
                                "Thursday 26 March 2026  05:43:39 +0000 (0:00:02.419)       0:41:02.672 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set num_osds] ************************************************************",
                                "Thursday 26 March 2026  05:43:40 +0000 (0:00:01.127)       0:41:03.800 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set_fact container_exec_cmd_osd] *****************************************",
                                "Thursday 26 March 2026  05:43:40 +0000 (0:00:00.790)       0:41:04.591 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Stop ceph osd] ***********************************************************",
                                "Thursday 26 March 2026  05:43:41 +0000 (0:00:00.905)       0:41:05.496 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:43:45 +0000 (0:00:03.814)       0:41:09.311 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:43:46 +0000 (0:00:01.107)       0:41:10.419 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:43:47 +0000 (0:00:01.151)       0:41:11.570 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:43:49 +0000 (0:00:01.134)       0:41:12.704 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:43:50 +0000 (0:00:01.506)       0:41:14.211 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:43:52 +0000 (0:00:01.549)       0:41:15.761 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:43:53 +0000 (0:00:01.497)       0:41:17.258 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:43:54 +0000 (0:00:01.149)       0:41:18.408 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:43:55 +0000 (0:00:01.151)       0:41:19.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:43:57 +0000 (0:00:01.108)       0:41:20.668 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:43:58 +0000 (0:00:01.534)       0:41:22.202 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:44:00 +0000 (0:00:01.608)       0:41:23.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:44:00 +0000 (0:00:00.750)       0:41:24.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:44:01 +0000 (0:00:00.796)       0:41:25.359 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:44:02 +0000 (0:00:00.812)       0:41:26.171 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:44:03 +0000 (0:00:00.775)       0:41:26.947 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:44:04 +0000 (0:00:00.816)       0:41:27.763 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:44:04 +0000 (0:00:00.812)       0:41:28.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:44:05 +0000 (0:00:00.765)       0:41:29.341 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:44:06 +0000 (0:00:00.781)       0:41:30.123 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:44:07 +0000 (0:00:00.786)       0:41:30.910 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:44:08 +0000 (0:00:00.939)       0:41:31.849 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:44:09 +0000 (0:00:00.812)       0:41:32.662 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:44:09 +0000 (0:00:00.792)       0:41:33.454 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:44:10 +0000 (0:00:00.784)       0:41:34.239 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:44:11 +0000 (0:00:00.781)       0:41:35.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:44:12 +0000 (0:00:00.790)       0:41:35.810 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:44:12 +0000 (0:00:00.798)       0:41:36.609 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:44:13 +0000 (0:00:00.754)       0:41:37.363 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:44:14 +0000 (0:00:00.772)       0:41:38.136 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:44:15 +0000 (0:00:00.825)       0:41:38.962 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:44:16 +0000 (0:00:00.790)       0:41:39.752 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:44:16 +0000 (0:00:00.762)       0:41:40.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:44:17 +0000 (0:00:00.931)       0:41:41.447 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:44:19 +0000 (0:00:01.609)       0:41:43.057 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:44:21 +0000 (0:00:01.837)       0:41:44.894 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:44:22 +0000 (0:00:01.157)       0:41:46.052 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:44:23 +0000 (0:00:01.185)       0:41:47.237 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:44:24 +0000 (0:00:01.156)       0:41:48.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:44:26 +0000 (0:00:01.772)       0:41:50.167 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:44:27 +0000 (0:00:01.473)       0:41:51.640 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:44:29 +0000 (0:00:01.153)       0:41:52.793 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:44:29 +0000 (0:00:00.785)       0:41:53.580 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:44:30 +0000 (0:00:00.785)       0:41:54.365 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:44:32 +0000 (0:00:01.379)       0:41:55.744 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:44:33 +0000 (0:00:01.802)       0:41:57.547 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:44:35 +0000 (0:00:01.138)       0:41:58.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:44:36 +0000 (0:00:01.134)       0:41:59.820 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:44:37 +0000 (0:00:01.165)       0:42:00.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:44:38 +0000 (0:00:01.130)       0:42:02.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:44:39 +0000 (0:00:01.184)       0:42:03.301 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:44:40 +0000 (0:00:00.798)       0:42:04.099 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:44:42 +0000 (0:00:02.151)       0:42:06.251 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:44:43 +0000 (0:00:00.758)       0:42:07.009 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:44:44 +0000 (0:00:01.118)       0:42:08.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:44:45 +0000 (0:00:01.266)       0:42:09.394 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:44:46 +0000 (0:00:01.150)       0:42:10.545 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:44:48 +0000 (0:00:01.157)       0:42:11.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:44:49 +0000 (0:00:01.124)       0:42:12.827 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:44:50 +0000 (0:00:01.147)       0:42:13.974 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:44:51 +0000 (0:00:01.118)       0:42:15.093 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:44:52 +0000 (0:00:01.139)       0:42:16.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:44:53 +0000 (0:00:01.123)       0:42:17.356 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:44:54 +0000 (0:00:00.834)       0:42:18.191 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:44:55 +0000 (0:00:01.149)       0:42:19.340 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:45:01 +0000 (0:00:06.153)       0:42:25.494 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  05:45:03 +0000 (0:00:01.273)       0:42:26.768 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  05:45:04 +0000 (0:00:01.488)       0:42:28.257 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:45:06 +0000 (0:00:01.616)       0:42:29.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:45:06 +0000 (0:00:00.734)       0:42:30.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:45:07 +0000 (0:00:00.749)       0:42:31.357 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:45:08 +0000 (0:00:00.770)       0:42:32.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:45:09 +0000 (0:00:00.838)       0:42:32.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:45:10 +0000 (0:00:00.737)       0:42:33.704 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:45:10 +0000 (0:00:00.789)       0:42:34.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:45:11 +0000 (0:00:00.757)       0:42:35.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:45:12 +0000 (0:00:00.745)       0:42:35.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:45:13 +0000 (0:00:00.787)       0:42:36.784 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:45:13 +0000 (0:00:00.779)       0:42:37.563 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:45:14 +0000 (0:00:00.912)       0:42:38.476 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:45:18 +0000 (0:00:04.172)       0:42:42.648 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:45:19 +0000 (0:00:00.850)       0:42:43.498 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:45:27 +0000 (0:00:07.704)       0:42:51.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:45:28 +0000 (0:00:00.773)       0:42:51.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:45:29 +0000 (0:00:00.853)       0:42:52.829 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:45:29 +0000 (0:00:00.816)       0:42:53.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:45:30 +0000 (0:00:00.774)       0:42:54.421 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:45:31 +0000 (0:00:00.813)       0:42:55.235 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:45:32 +0000 (0:00:00.911)       0:42:56.147 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:45:33 +0000 (0:00:01.106)       0:42:57.253 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:45:35 +0000 (0:00:01.529)       0:42:58.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:45:36 +0000 (0:00:01.469)       0:43:00.252 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:45:37 +0000 (0:00:00.875)       0:43:01.128 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:45:38 +0000 (0:00:00.983)       0:43:02.111 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Thursday 26 March 2026  05:45:39 +0000 (0:00:01.388)       0:43:03.500 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:45:40 +0000 (0:00:00.778)       0:43:04.279 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Thursday 26 March 2026  05:45:41 +0000 (0:00:01.361)       0:43:05.640 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Thursday 26 March 2026  05:45:43 +0000 (0:00:01.123)       0:43:06.764 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Thursday 26 March 2026  05:45:44 +0000 (0:00:01.139)       0:43:07.904 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Thursday 26 March 2026  05:45:45 +0000 (0:00:01.177)       0:43:09.081 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Thursday 26 March 2026  05:45:46 +0000 (0:00:01.440)       0:43:10.522 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Thursday 26 March 2026  05:45:48 +0000 (0:00:01.161)       0:43:11.684 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Thursday 26 March 2026  05:45:50 +0000 (0:00:02.610)       0:43:14.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Thursday 26 March 2026  05:45:51 +0000 (0:00:00.776)       0:43:15.071 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Thursday 26 March 2026  05:45:52 +0000 (0:00:01.148)       0:43:16.220 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:45:54 +0000 (0:00:01.848)       0:43:18.068 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:45:57 +0000 (0:00:03.191)       0:43:21.260 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Thursday 26 March 2026  05:45:59 +0000 (0:00:01.582)       0:43:22.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Thursday 26 March 2026  05:46:00 +0000 (0:00:00.865)       0:43:23.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Thursday 26 March 2026  05:46:00 +0000 (0:00:00.802)       0:43:24.511 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Thursday 26 March 2026  05:46:01 +0000 (0:00:00.763)       0:43:25.275 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Thursday 26 March 2026  05:46:02 +0000 (0:00:01.133)       0:43:26.409 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Thursday 26 March 2026  05:46:04 +0000 (0:00:01.473)       0:43:27.882 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:46:07 +0000 (0:00:03.291)       0:43:31.174 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:46:08 +0000 (0:00:01.241)       0:43:32.416 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Thursday 26 March 2026  05:46:10 +0000 (0:00:01.995)       0:43:34.412 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Thursday 26 March 2026  05:46:12 +0000 (0:00:01.938)       0:43:36.350 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:46:14 +0000 (0:00:02.246)       0:43:38.597 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:46:16 +0000 (0:00:01.201)       0:43:39.799 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Thursday 26 March 2026  05:46:17 +0000 (0:00:01.182)       0:43:40.981 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Thursday 26 March 2026  05:46:19 +0000 (0:00:01.812)       0:43:42.793 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Thursday 26 March 2026  05:46:22 +0000 (0:00:02.900)       0:43:45.694 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Thursday 26 March 2026  05:46:26 +0000 (0:00:04.206)       0:43:49.900 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Thursday 26 March 2026  05:46:27 +0000 (0:00:00.889)       0:43:50.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Thursday 26 March 2026  05:46:28 +0000 (0:00:00.885)       0:43:51.676 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************",
                                "Thursday 26 March 2026  05:46:29 +0000 (0:00:01.016)       0:43:52.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***",
                                "Thursday 26 March 2026  05:46:29 +0000 (0:00:00.750)       0:43:53.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Waiting for clean pgs...] ************************************************",
                                "Thursday 26 March 2026  05:46:30 +0000 (0:00:00.789)       0:43:54.232 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (600 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (599 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (598 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for clean pgs... (597 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:46:44 +0000 (0:00:13.720)       0:44:07.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:46:45 +0000 (0:00:00.916)       0:44:08.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:46:46 +0000 (0:00:00.788)       0:44:09.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:46:46 +0000 (0:00:00.767)       0:44:10.426 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:46:47 +0000 (0:00:00.770)       0:44:11.197 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:46:48 +0000 (0:00:00.788)       0:44:11.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:46:49 +0000 (0:00:00.769)       0:44:12.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph osds cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:46:50 +0000 (0:00:00.960)       0:44:13.715 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:46:51 +0000 (0:00:01.338)       0:44:15.053 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:46:52 +0000 (0:00:01.454)       0:44:16.508 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:46:53 +0000 (0:00:01.108)       0:44:17.617 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:46:55 +0000 (0:00:01.485)       0:44:19.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:46:56 +0000 (0:00:01.191)       0:44:20.294 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:46:57 +0000 (0:00:01.128)       0:44:21.423 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:46:58 +0000 (0:00:01.167)       0:44:22.590 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:47:00 +0000 (0:00:01.132)       0:44:23.723 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:47:01 +0000 (0:00:01.148)       0:44:24.871 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:47:03 +0000 (0:00:02.057)       0:44:26.929 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:47:04 +0000 (0:00:01.237)       0:44:28.166 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:47:07 +0000 (0:00:03.293)       0:44:31.460 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:47:09 +0000 (0:00:01.791)       0:44:33.252 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:47:11 +0000 (0:00:01.620)       0:44:34.872 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:47:12 +0000 (0:00:01.214)       0:44:36.087 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:47:05.394004', 'end': '2026-03-26 05:47:05.430712', 'delta': '0:00:00.036708', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:47:05.989006', 'end': '2026-03-26 05:47:06.037018', 'delta': '0:00:00.048012', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:47:06.596719', 'end': '2026-03-26 05:47:06.655213', 'delta': '0:00:00.058494', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:47:13 +0000 (0:00:01.178)       0:44:37.265 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:47:14 +0000 (0:00:01.258)       0:44:38.524 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:47:16 +0000 (0:00:01.293)       0:44:39.818 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:47:17 +0000 (0:00:01.139)       0:44:40.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:47:19 +0000 (0:00:02.028)       0:44:42.986 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:47:20 +0000 (0:00:01.319)       0:44:44.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:47:21 +0000 (0:00:01.155)       0:44:45.462 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:47:23 +0000 (0:00:01.251)       0:44:46.713 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:47:24 +0000 (0:00:01.141)       0:44:47.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:47:25 +0000 (0:00:01.154)       0:44:49.009 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:47:26 +0000 (0:00:01.234)       0:44:50.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:47:27 +0000 (0:00:01.124)       0:44:51.368 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:47:28 +0000 (0:00:01.176)       0:44:52.544 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:47:30 +0000 (0:00:01.146)       0:44:53.691 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:47:31 +0000 (0:00:01.186)       0:44:54.878 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:47:32 +0000 (0:00:01.422)       0:44:56.301 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:47:34 +0000 (0:00:01.415)       0:44:57.717 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:47:36 +0000 (0:00:02.516)       0:45:00.233 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:47:37 +0000 (0:00:01.138)       0:45:01.372 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:47:39 +0000 (0:00:01.502)       0:45:02.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:47:40 +0000 (0:00:01.151)       0:45:04.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:47:41 +0000 (0:00:01.240)       0:45:05.267 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:47:42 +0000 (0:00:01.173)       0:45:06.440 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:47:44 +0000 (0:00:02.128)       0:45:08.569 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:47:46 +0000 (0:00:01.204)       0:45:09.774 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:47:47 +0000 (0:00:01.160)       0:45:10.934 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:47:48 +0000 (0:00:01.141)       0:45:12.076 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:47:49 +0000 (0:00:01.115)       0:45:13.192 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:47:50 +0000 (0:00:01.168)       0:45:14.361 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:47:51 +0000 (0:00:01.257)       0:45:15.618 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:47:53 +0000 (0:00:01.387)       0:45:17.005 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:47:54 +0000 (0:00:01.362)       0:45:18.368 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:47:56 +0000 (0:00:01.417)       0:45:19.786 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:47:57 +0000 (0:00:01.184)       0:45:20.970 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:47:59 +0000 (0:00:01.756)       0:45:22.727 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:48:01 +0000 (0:00:02.170)       0:45:24.898 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Get osd numbers - non container] *****************************************",
                                "Thursday 26 March 2026  05:48:03 +0000 (0:00:02.264)       0:45:27.162 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set num_osds] ************************************************************",
                                "Thursday 26 March 2026  05:48:04 +0000 (0:00:01.150)       0:45:28.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set_fact container_exec_cmd_osd] *****************************************",
                                "Thursday 26 March 2026  05:48:05 +0000 (0:00:00.779)       0:45:29.092 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Stop ceph osd] ***********************************************************",
                                "Thursday 26 March 2026  05:48:06 +0000 (0:00:00.850)       0:45:29.943 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:48:10 +0000 (0:00:04.686)       0:45:34.630 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:48:12 +0000 (0:00:01.150)       0:45:35.780 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:48:13 +0000 (0:00:01.098)       0:45:36.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:48:14 +0000 (0:00:01.112)       0:45:37.991 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:48:15 +0000 (0:00:01.503)       0:45:39.495 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:48:17 +0000 (0:00:01.614)       0:45:41.110 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:48:18 +0000 (0:00:01.536)       0:45:42.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:48:20 +0000 (0:00:01.113)       0:45:43.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:48:21 +0000 (0:00:01.121)       0:45:44.882 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:48:22 +0000 (0:00:01.131)       0:45:46.014 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:48:23 +0000 (0:00:01.518)       0:45:47.533 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:48:25 +0000 (0:00:01.632)       0:45:49.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:48:26 +0000 (0:00:00.787)       0:45:49.952 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:48:27 +0000 (0:00:00.793)       0:45:50.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:48:27 +0000 (0:00:00.779)       0:45:51.525 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:48:28 +0000 (0:00:00.835)       0:45:52.361 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:48:29 +0000 (0:00:00.779)       0:45:53.140 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:48:30 +0000 (0:00:00.744)       0:45:53.885 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:48:30 +0000 (0:00:00.762)       0:45:54.647 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:48:31 +0000 (0:00:00.816)       0:45:55.463 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:48:32 +0000 (0:00:00.819)       0:45:56.282 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:48:33 +0000 (0:00:00.790)       0:45:57.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:48:34 +0000 (0:00:00.769)       0:45:57.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:48:34 +0000 (0:00:00.768)       0:45:58.611 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:48:35 +0000 (0:00:00.761)       0:45:59.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:48:36 +0000 (0:00:00.770)       0:46:00.144 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:48:37 +0000 (0:00:00.783)       0:46:00.928 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:48:38 +0000 (0:00:00.769)       0:46:01.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:48:38 +0000 (0:00:00.762)       0:46:02.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:48:39 +0000 (0:00:00.783)       0:46:03.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:48:40 +0000 (0:00:00.760)       0:46:04.005 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:48:41 +0000 (0:00:00.915)       0:46:04.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:48:42 +0000 (0:00:00.794)       0:46:05.715 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:48:42 +0000 (0:00:00.747)       0:46:06.463 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:48:44 +0000 (0:00:01.632)       0:46:08.095 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:48:46 +0000 (0:00:01.902)       0:46:09.998 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:48:47 +0000 (0:00:01.129)       0:46:11.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:48:48 +0000 (0:00:01.129)       0:46:12.256 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:48:49 +0000 (0:00:01.132)       0:46:13.389 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:48:51 +0000 (0:00:01.803)       0:46:15.193 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:48:53 +0000 (0:00:01.466)       0:46:16.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:48:54 +0000 (0:00:01.114)       0:46:17.774 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:48:55 +0000 (0:00:00.894)       0:46:18.668 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:48:55 +0000 (0:00:00.807)       0:46:19.476 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:48:56 +0000 (0:00:01.157)       0:46:20.633 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:48:58 +0000 (0:00:01.828)       0:46:22.463 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:48:59 +0000 (0:00:01.137)       0:46:23.601 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:49:01 +0000 (0:00:01.164)       0:46:24.765 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:49:02 +0000 (0:00:01.166)       0:46:25.932 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:49:03 +0000 (0:00:01.132)       0:46:27.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:49:04 +0000 (0:00:01.165)       0:46:28.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:49:05 +0000 (0:00:00.788)       0:46:29.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:49:07 +0000 (0:00:02.070)       0:46:31.088 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:49:08 +0000 (0:00:00.772)       0:46:31.861 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:49:09 +0000 (0:00:01.276)       0:46:33.137 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:49:10 +0000 (0:00:01.178)       0:46:34.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:49:11 +0000 (0:00:01.136)       0:46:35.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:49:12 +0000 (0:00:01.155)       0:46:36.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:49:14 +0000 (0:00:01.147)       0:46:37.756 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:49:15 +0000 (0:00:01.265)       0:46:39.021 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:49:16 +0000 (0:00:01.162)       0:46:40.184 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:49:17 +0000 (0:00:01.115)       0:46:41.300 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:49:18 +0000 (0:00:01.140)       0:46:42.440 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:49:19 +0000 (0:00:00.804)       0:46:43.244 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:49:20 +0000 (0:00:01.109)       0:46:44.354 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:49:27 +0000 (0:00:06.443)       0:46:50.798 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  05:49:28 +0000 (0:00:01.159)       0:46:51.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  05:49:29 +0000 (0:00:01.498)       0:46:53.456 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:49:31 +0000 (0:00:01.587)       0:46:55.044 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:49:32 +0000 (0:00:00.838)       0:46:55.883 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:49:33 +0000 (0:00:00.797)       0:46:56.681 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:49:33 +0000 (0:00:00.763)       0:46:57.444 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:49:34 +0000 (0:00:00.798)       0:46:58.242 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:49:35 +0000 (0:00:00.764)       0:46:59.007 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:49:36 +0000 (0:00:00.757)       0:46:59.765 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:49:36 +0000 (0:00:00.750)       0:47:00.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:49:37 +0000 (0:00:00.771)       0:47:01.287 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:49:38 +0000 (0:00:00.840)       0:47:02.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:49:39 +0000 (0:00:00.786)       0:47:02.915 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:49:40 +0000 (0:00:00.848)       0:47:03.763 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:49:44 +0000 (0:00:04.124)       0:47:07.888 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:49:45 +0000 (0:00:00.850)       0:47:08.738 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:49:52 +0000 (0:00:07.356)       0:47:16.095 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:49:53 +0000 (0:00:00.790)       0:47:16.885 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:49:54 +0000 (0:00:00.790)       0:47:17.676 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:49:54 +0000 (0:00:00.824)       0:47:18.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:49:55 +0000 (0:00:00.816)       0:47:19.318 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:49:56 +0000 (0:00:00.794)       0:47:20.113 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:49:57 +0000 (0:00:00.932)       0:47:21.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:49:58 +0000 (0:00:01.433)       0:47:22.478 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:50:00 +0000 (0:00:01.487)       0:47:23.966 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:50:01 +0000 (0:00:01.081)       0:47:25.048 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:50:02 +0000 (0:00:00.805)       0:47:25.854 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:50:03 +0000 (0:00:01.035)       0:47:26.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Thursday 26 March 2026  05:50:04 +0000 (0:00:01.481)       0:47:28.370 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Thursday 26 March 2026  05:50:05 +0000 (0:00:00.794)       0:47:29.165 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Thursday 26 March 2026  05:50:06 +0000 (0:00:01.470)       0:47:30.635 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Thursday 26 March 2026  05:50:08 +0000 (0:00:01.062)       0:47:31.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Thursday 26 March 2026  05:50:09 +0000 (0:00:01.124)       0:47:32.822 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Thursday 26 March 2026  05:50:10 +0000 (0:00:01.083)       0:47:33.905 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Thursday 26 March 2026  05:50:11 +0000 (0:00:01.468)       0:47:35.374 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Thursday 26 March 2026  05:50:12 +0000 (0:00:01.108)       0:47:36.482 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Thursday 26 March 2026  05:50:16 +0000 (0:00:03.633)       0:47:40.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Thursday 26 March 2026  05:50:17 +0000 (0:00:00.765)       0:47:40.882 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Thursday 26 March 2026  05:50:18 +0000 (0:00:01.113)       0:47:41.995 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:50:20 +0000 (0:00:01.798)       0:47:43.794 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:50:23 +0000 (0:00:03.176)       0:47:46.971 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Thursday 26 March 2026  05:50:24 +0000 (0:00:01.623)       0:47:48.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Thursday 26 March 2026  05:50:25 +0000 (0:00:00.906)       0:47:49.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Thursday 26 March 2026  05:50:26 +0000 (0:00:00.770)       0:47:50.271 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Thursday 26 March 2026  05:50:27 +0000 (0:00:00.748)       0:47:51.020 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Thursday 26 March 2026  05:50:28 +0000 (0:00:01.235)       0:47:52.255 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Thursday 26 March 2026  05:50:30 +0000 (0:00:01.465)       0:47:53.721 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:50:33 +0000 (0:00:03.351)       0:47:57.072 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:50:34 +0000 (0:00:01.106)       0:47:58.179 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Thursday 26 March 2026  05:50:36 +0000 (0:00:01.966)       0:48:00.145 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Thursday 26 March 2026  05:50:38 +0000 (0:00:01.927)       0:48:02.073 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Thursday 26 March 2026  05:50:40 +0000 (0:00:02.255)       0:48:04.329 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Thursday 26 March 2026  05:50:41 +0000 (0:00:01.120)       0:48:05.449 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Thursday 26 March 2026  05:50:42 +0000 (0:00:01.154)       0:48:06.604 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Thursday 26 March 2026  05:50:44 +0000 (0:00:01.848)       0:48:08.452 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Thursday 26 March 2026  05:50:47 +0000 (0:00:02.863)       0:48:11.315 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Thursday 26 March 2026  05:50:51 +0000 (0:00:04.175)       0:48:15.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Thursday 26 March 2026  05:50:52 +0000 (0:00:00.909)       0:48:16.401 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Thursday 26 March 2026  05:51:06 +0000 (0:00:13.348)       0:48:29.749 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Scan ceph-disk osds with ceph-volume if deploying nautilus] **************",
                                "Thursday 26 March 2026  05:51:06 +0000 (0:00:00.882)       0:48:30.632 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Activate scanned ceph-disk osds and migrate to ceph-volume if deploying nautilus] ***",
                                "Thursday 26 March 2026  05:51:07 +0000 (0:00:00.749)       0:48:31.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Waiting for clean pgs...] ************************************************",
                                "Thursday 26 March 2026  05:51:08 +0000 (0:00:00.787)       0:48:32.169 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Thursday 26 March 2026  05:51:10 +0000 (0:00:01.931)       0:48:34.101 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Thursday 26 March 2026  05:51:11 +0000 (0:00:00.779)       0:48:34.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Thursday 26 March 2026  05:51:12 +0000 (0:00:00.814)       0:48:35.695 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Thursday 26 March 2026  05:51:12 +0000 (0:00:00.795)       0:48:36.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Thursday 26 March 2026  05:51:13 +0000 (0:00:00.776)       0:48:37.268 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Thursday 26 March 2026  05:51:14 +0000 (0:00:00.769)       0:48:38.037 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Thursday 26 March 2026  05:51:15 +0000 (0:00:00.764)       0:48:38.802 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Complete osd upgrade] ****************************************************",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:51:16 +0000 (0:00:01.800)       0:48:40.602 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:51:18 +0000 (0:00:01.868)       0:48:42.471 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Re-enable pg autoscale on pools] *****************************************",
                                "Thursday 26 March 2026  05:51:20 +0000 (0:00:01.358)       0:48:43.830 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.mgr', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_data', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'cephfs_metadata', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.data', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.buckets.index', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.control', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.log', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': 'default.rgw.meta', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'name': '.rgw.root', 'mode': 'on'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'backups', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'volumes', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'images', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'metrics', 'mode': 'off'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vms', 'mode': 'off'}) \u001b[0m",
                                "",
                                "TASK [Unset osd flags] *********************************************************",
                                "Thursday 26 March 2026  05:52:34 +0000 (0:01:14.512)       0:49:58.342 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=noout)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=nodeep-scrub)\u001b[0m",
                                "",
                                "TASK [Re-enable balancer] ******************************************************",
                                "Thursday 26 March 2026  05:52:41 +0000 (0:00:06.923)       0:50:05.265 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph mdss cluster, deactivate all rank > 0] **********************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:52:44 +0000 (0:00:03.174)       0:50:08.439 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:52:45 +0000 (0:00:01.181)       0:50:09.621 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:52:47 +0000 (0:00:01.579)       0:50:11.201 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:52:48 +0000 (0:00:01.207)       0:50:12.409 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:52:50 +0000 (0:00:01.523)       0:50:13.932 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:52:51 +0000 (0:00:01.180)       0:50:15.113 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:52:52 +0000 (0:00:01.197)       0:50:16.310 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:52:53 +0000 (0:00:01.188)       0:50:17.499 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:52:55 +0000 (0:00:01.166)       0:50:18.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:52:56 +0000 (0:00:01.243)       0:50:19.909 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:52:57 +0000 (0:00:01.682)       0:50:21.592 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:52:59 +0000 (0:00:01.300)       0:50:22.893 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:53:02 +0000 (0:00:03.160)       0:50:26.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:53:03 +0000 (0:00:01.490)       0:50:27.543 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:53:05 +0000 (0:00:01.726)       0:50:29.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:53:06 +0000 (0:00:01.215)       0:50:30.486 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:52:59.842534', 'end': '2026-03-26 05:52:59.895695', 'delta': '0:00:00.053161', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:53:00.541458', 'end': '2026-03-26 05:53:00.583420', 'delta': '0:00:00.041962', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:53:01.105903', 'end': '2026-03-26 05:53:01.153302', 'delta': '0:00:00.047399', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:53:08 +0000 (0:00:01.263)       0:50:31.749 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:53:09 +0000 (0:00:01.304)       0:50:33.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:53:10 +0000 (0:00:01.316)       0:50:34.371 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:53:11 +0000 (0:00:01.155)       0:50:35.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:53:13 +0000 (0:00:02.035)       0:50:37.561 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:53:15 +0000 (0:00:01.161)       0:50:38.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:53:16 +0000 (0:00:01.250)       0:50:39.974 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:53:17 +0000 (0:00:01.273)       0:50:41.247 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:53:18 +0000 (0:00:01.155)       0:50:42.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:53:19 +0000 (0:00:01.147)       0:50:43.550 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:53:21 +0000 (0:00:01.180)       0:50:44.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:53:22 +0000 (0:00:01.190)       0:50:45.921 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:53:23 +0000 (0:00:01.132)       0:50:47.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:53:24 +0000 (0:00:01.180)       0:50:48.235 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:53:25 +0000 (0:00:01.166)       0:50:49.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:53:27 +0000 (0:00:01.310)       0:50:50.712 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-12-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'c374eb4c', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1', 'scsi-SQEMU_QEMU_HARDDISK_c374eb4c-3572-4f0b-927c-38d35765f44a-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:53:28 +0000 (0:00:01.342)       0:50:52.055 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:53:29 +0000 (0:00:01.543)       0:50:53.599 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:53:31 +0000 (0:00:01.182)       0:50:54.781 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:53:32 +0000 (0:00:01.520)       0:50:56.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:53:33 +0000 (0:00:01.181)       0:50:57.483 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:53:35 +0000 (0:00:01.248)       0:50:58.732 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:53:36 +0000 (0:00:01.141)       0:50:59.873 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:53:37 +0000 (0:00:01.714)       0:51:01.588 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:53:39 +0000 (0:00:01.198)       0:51:02.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:53:40 +0000 (0:00:01.157)       0:51:03.944 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:53:42 +0000 (0:00:02.213)       0:51:06.157 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Set max_mds 1 on ceph fs] ************************************************",
                                "Thursday 26 March 2026  05:53:45 +0000 (0:00:02.776)       0:51:08.933 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Wait until only rank 0 is up] ********************************************",
                                "Thursday 26 March 2026  05:53:48 +0000 (0:00:03.190)       0:51:12.124 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get name of remaining active mds] ****************************************",
                                "Thursday 26 March 2026  05:53:51 +0000 (0:00:02.985)       0:51:15.110 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set_fact mds_active_name] ************************************************",
                                "Thursday 26 March 2026  05:53:53 +0000 (0:00:02.115)       0:51:17.225 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'gid_4752', 'value': {'gid': 4752, 'name': 'testbed-node-3', 'rank': 0, 'incarnation': 4, 'state': 'up:active', 'state_seq': 2, 'addr': '192.168.16.13:6817/3459244977', 'addrs': {'addrvec': [{'type': 'v2', 'addr': '192.168.16.13:6816', 'nonce': 3459244977}, {'type': 'v1', 'addr': '192.168.16.13:6817', 'nonce': 3459244977}]}, 'join_fscid': -1, 'export_targets': [], 'features': 4540138322906710015, 'flags': 0, 'compat': {'compat': {}, 'ro_compat': {}, 'incompat': {'feature_1': 'base v0.20', 'feature_2': 'client writeable ranges', 'feature_3': 'default file layouts on dirs', 'feature_4': 'dir inode in separate object', 'feature_5': 'mds uses versioned encoding', 'feature_6': 'dirfrag is stored in omap', 'feature_7': 'mds uses inline data', 'feature_8': 'no anchor table', 'feature_9': 'file layout v2', 'feature_10': 'snaprealm v2'}}}})\u001b[0m",
                                "",
                                "TASK [Set_fact mds_active_host] ************************************************",
                                "Thursday 26 March 2026  05:53:54 +0000 (0:00:01.240)       0:51:18.466 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "",
                                "TASK [Create standby_mdss group] ***********************************************",
                                "Thursday 26 March 2026  05:53:56 +0000 (0:00:01.553)       0:51:20.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [Stop standby ceph mds] ***************************************************",
                                "Thursday 26 March 2026  05:53:57 +0000 (0:00:01.559)       0:51:21.579 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [Mask systemd units for standby ceph mds] *********************************",
                                "Thursday 26 March 2026  05:54:09 +0000 (0:00:11.167)       0:51:32.747 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [Wait until all standbys mds are stopped] *********************************",
                                "Thursday 26 March 2026  05:54:13 +0000 (0:00:04.755)       0:51:37.502 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Create active_mdss group] ************************************************",
                                "Thursday 26 March 2026  05:54:16 +0000 (0:00:02.186)       0:51:39.688 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Upgrade active mds] ******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:54:17 +0000 (0:00:01.601)       0:51:41.290 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:54:18 +0000 (0:00:01.144)       0:51:42.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:54:20 +0000 (0:00:01.438)       0:51:43.873 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:54:21 +0000 (0:00:01.110)       0:51:44.983 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:54:22 +0000 (0:00:01.428)       0:51:46.412 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:54:23 +0000 (0:00:01.135)       0:51:47.548 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:54:25 +0000 (0:00:01.113)       0:51:48.661 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:54:26 +0000 (0:00:01.181)       0:51:49.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:54:27 +0000 (0:00:01.134)       0:51:50.977 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:54:28 +0000 (0:00:01.129)       0:51:52.107 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:54:30 +0000 (0:00:01.751)       0:51:53.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:54:31 +0000 (0:00:01.273)       0:51:55.131 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:54:34 +0000 (0:00:02.843)       0:51:57.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:54:35 +0000 (0:00:01.507)       0:51:59.482 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:54:37 +0000 (0:00:01.981)       0:52:01.464 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:54:38 +0000 (0:00:01.153)       0:52:02.618 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:54:32.014467', 'end': '2026-03-26 05:54:32.067837', 'delta': '0:00:00.053370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:54:32.555012', 'end': '2026-03-26 05:54:32.590382', 'delta': '0:00:00.035370', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:54:33.112871', 'end': '2026-03-26 05:54:33.160580', 'delta': '0:00:00.047709', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:54:40 +0000 (0:00:01.208)       0:52:03.826 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:54:41 +0000 (0:00:01.304)       0:52:05.131 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:54:43 +0000 (0:00:01.662)       0:52:06.793 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:54:44 +0000 (0:00:01.149)       0:52:07.942 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:54:46 +0000 (0:00:02.011)       0:52:09.954 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:54:47 +0000 (0:00:01.248)       0:52:11.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:54:48 +0000 (0:00:01.161)       0:52:12.365 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:54:49 +0000 (0:00:01.263)       0:52:13.629 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:54:51 +0000 (0:00:01.166)       0:52:14.795 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:54:52 +0000 (0:00:01.125)       0:52:15.921 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:54:53 +0000 (0:00:01.166)       0:52:17.087 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:54:54 +0000 (0:00:01.114)       0:52:18.201 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:54:55 +0000 (0:00:01.226)       0:52:19.428 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:54:56 +0000 (0:00:01.129)       0:52:20.557 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:54:58 +0000 (0:00:01.171)       0:52:21.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:54:59 +0000 (0:00:01.516)       0:52:23.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:55:01 +0000 (0:00:01.431)       0:52:24.678 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:55:02 +0000 (0:00:01.579)       0:52:26.257 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:55:03 +0000 (0:00:01.132)       0:52:27.390 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:55:05 +0000 (0:00:01.478)       0:52:28.869 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:55:06 +0000 (0:00:01.211)       0:52:30.081 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:55:07 +0000 (0:00:01.192)       0:52:31.273 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:55:08 +0000 (0:00:01.107)       0:52:32.381 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:55:10 +0000 (0:00:01.840)       0:52:34.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:55:11 +0000 (0:00:01.134)       0:52:35.356 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:55:12 +0000 (0:00:01.111)       0:52:36.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:55:13 +0000 (0:00:01.174)       0:52:37.642 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:55:15 +0000 (0:00:01.347)       0:52:38.990 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:55:16 +0000 (0:00:01.161)       0:52:40.152 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:55:17 +0000 (0:00:01.327)       0:52:41.479 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:55:19 +0000 (0:00:01.450)       0:52:42.930 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:55:20 +0000 (0:00:01.404)       0:52:44.335 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:55:22 +0000 (0:00:01.468)       0:52:45.803 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:55:23 +0000 (0:00:01.194)       0:52:46.998 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:55:24 +0000 (0:00:01.338)       0:52:48.336 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:55:26 +0000 (0:00:02.312)       0:52:50.649 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Prevent restart from the packaging] **************************************",
                                "Thursday 26 March 2026  05:55:29 +0000 (0:00:02.663)       0:52:53.312 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:55:30 +0000 (0:00:01.132)       0:52:54.445 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:55:31 +0000 (0:00:01.144)       0:52:55.589 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:55:33 +0000 (0:00:01.308)       0:52:56.898 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:55:34 +0000 (0:00:01.127)       0:52:58.025 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:55:35 +0000 (0:00:01.504)       0:52:59.530 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:55:37 +0000 (0:00:01.562)       0:53:01.093 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:55:38 +0000 (0:00:01.552)       0:53:02.646 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:55:40 +0000 (0:00:01.131)       0:53:03.778 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:55:41 +0000 (0:00:01.194)       0:53:04.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:55:42 +0000 (0:00:01.121)       0:53:06.094 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:55:44 +0000 (0:00:01.588)       0:53:07.682 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:55:45 +0000 (0:00:01.590)       0:53:09.272 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:55:46 +0000 (0:00:01.158)       0:53:10.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:55:47 +0000 (0:00:01.226)       0:53:11.657 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:55:49 +0000 (0:00:01.186)       0:53:12.844 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:55:50 +0000 (0:00:01.126)       0:53:13.971 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:55:51 +0000 (0:00:01.182)       0:53:15.153 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:55:52 +0000 (0:00:01.164)       0:53:16.317 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  05:55:53 +0000 (0:00:01.201)       0:53:17.518 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  05:55:55 +0000 (0:00:01.159)       0:53:18.678 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  05:55:56 +0000 (0:00:01.157)       0:53:19.836 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  05:55:57 +0000 (0:00:01.175)       0:53:21.011 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  05:55:58 +0000 (0:00:01.145)       0:53:22.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  05:55:59 +0000 (0:00:01.151)       0:53:23.309 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  05:56:00 +0000 (0:00:01.158)       0:53:24.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  05:56:01 +0000 (0:00:01.132)       0:53:25.601 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  05:56:03 +0000 (0:00:01.237)       0:53:26.838 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  05:56:04 +0000 (0:00:01.145)       0:53:27.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  05:56:05 +0000 (0:00:01.191)       0:53:29.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  05:56:06 +0000 (0:00:01.141)       0:53:30.317 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  05:56:07 +0000 (0:00:01.202)       0:53:31.519 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  05:56:09 +0000 (0:00:01.180)       0:53:32.699 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  05:56:10 +0000 (0:00:01.135)       0:53:33.835 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  05:56:11 +0000 (0:00:01.128)       0:53:34.964 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  05:56:13 +0000 (0:00:01.889)       0:53:36.853 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  05:56:15 +0000 (0:00:02.250)       0:53:39.104 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  05:56:16 +0000 (0:00:01.120)       0:53:40.224 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  05:56:17 +0000 (0:00:01.154)       0:53:41.378 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  05:56:19 +0000 (0:00:01.304)       0:53:42.682 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  05:56:20 +0000 (0:00:01.869)       0:53:44.553 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  05:56:22 +0000 (0:00:01.497)       0:53:46.051 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  05:56:23 +0000 (0:00:01.144)       0:53:47.196 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  05:56:24 +0000 (0:00:01.155)       0:53:48.351 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  05:56:25 +0000 (0:00:01.180)       0:53:49.532 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  05:56:27 +0000 (0:00:01.190)       0:53:50.723 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  05:56:28 +0000 (0:00:01.815)       0:53:52.538 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  05:56:30 +0000 (0:00:01.166)       0:53:53.705 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  05:56:31 +0000 (0:00:01.166)       0:53:54.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  05:56:32 +0000 (0:00:01.187)       0:53:56.059 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  05:56:33 +0000 (0:00:01.199)       0:53:57.258 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  05:56:34 +0000 (0:00:01.391)       0:53:58.650 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  05:56:36 +0000 (0:00:01.196)       0:53:59.847 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  05:56:38 +0000 (0:00:02.550)       0:54:02.398 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  05:56:39 +0000 (0:00:01.135)       0:54:03.533 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  05:56:41 +0000 (0:00:01.155)       0:54:04.689 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  05:56:42 +0000 (0:00:01.175)       0:54:05.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  05:56:43 +0000 (0:00:01.150)       0:54:07.016 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  05:56:44 +0000 (0:00:01.146)       0:54:08.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  05:56:45 +0000 (0:00:01.101)       0:54:09.264 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  05:56:46 +0000 (0:00:01.177)       0:54:10.442 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  05:56:47 +0000 (0:00:01.186)       0:54:11.629 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  05:56:49 +0000 (0:00:01.124)       0:54:12.754 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  05:56:50 +0000 (0:00:01.228)       0:54:13.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  05:56:51 +0000 (0:00:01.151)       0:54:15.133 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  05:56:52 +0000 (0:00:01.127)       0:54:16.261 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  05:56:59 +0000 (0:00:06.553)       0:54:22.815 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  05:57:00 +0000 (0:00:01.161)       0:54:23.976 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  05:57:01 +0000 (0:00:01.498)       0:54:25.474 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  05:57:04 +0000 (0:00:03.000)       0:54:28.475 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  05:57:05 +0000 (0:00:01.122)       0:54:29.598 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  05:57:07 +0000 (0:00:01.118)       0:54:30.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  05:57:08 +0000 (0:00:01.153)       0:54:31.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  05:57:09 +0000 (0:00:01.214)       0:54:33.085 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  05:57:10 +0000 (0:00:01.139)       0:54:34.224 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  05:57:11 +0000 (0:00:01.126)       0:54:35.351 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  05:57:12 +0000 (0:00:01.113)       0:54:36.465 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  05:57:13 +0000 (0:00:01.106)       0:54:37.572 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  05:57:15 +0000 (0:00:01.128)       0:54:38.701 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  05:57:16 +0000 (0:00:01.172)       0:54:39.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  05:57:17 +0000 (0:00:01.211)       0:54:41.085 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  05:57:21 +0000 (0:00:04.492)       0:54:45.578 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  05:57:23 +0000 (0:00:01.201)       0:54:46.779 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  05:57:27 +0000 (0:00:04.801)       0:54:51.581 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  05:57:29 +0000 (0:00:01.133)       0:54:52.714 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:57:30 +0000 (0:00:01.110)       0:54:53.825 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:57:31 +0000 (0:00:01.222)       0:54:55.048 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:57:32 +0000 (0:00:01.180)       0:54:56.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:57:33 +0000 (0:00:01.184)       0:54:57.413 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:57:34 +0000 (0:00:01.233)       0:54:58.647 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:57:36 +0000 (0:00:01.520)       0:55:00.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:57:37 +0000 (0:00:01.450)       0:55:01.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:57:39 +0000 (0:00:01.470)       0:55:03.089 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:57:40 +0000 (0:00:01.175)       0:55:04.265 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  05:57:41 +0000 (0:00:01.384)       0:55:05.650 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Thursday 26 March 2026  05:57:43 +0000 (0:00:01.845)       0:55:07.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  05:57:44 +0000 (0:00:01.154)       0:55:08.649 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Thursday 26 March 2026  05:57:46 +0000 (0:00:01.819)       0:55:10.469 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  05:57:48 +0000 (0:00:01.987)       0:55:12.457 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  05:57:51 +0000 (0:00:03.181)       0:55:15.639 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Thursday 26 March 2026  05:57:53 +0000 (0:00:01.946)       0:55:17.585 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Thursday 26 March 2026  05:57:55 +0000 (0:00:01.579)       0:55:19.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Thursday 26 March 2026  05:57:56 +0000 (0:00:01.149)       0:55:20.314 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  05:57:58 +0000 (0:00:01.552)       0:55:21.867 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  05:57:59 +0000 (0:00:01.576)       0:55:23.444 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Thursday 26 March 2026  05:58:01 +0000 (0:00:01.996)       0:55:25.440 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Thursday 26 March 2026  05:58:03 +0000 (0:00:01.894)       0:55:27.335 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Thursday 26 March 2026  05:58:05 +0000 (0:00:02.290)       0:55:29.625 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Thursday 26 March 2026  05:58:08 +0000 (0:00:02.324)       0:55:31.950 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Restart ceph mds] ********************************************************",
                                "Thursday 26 March 2026  05:58:09 +0000 (0:00:01.635)       0:55:33.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Restart active mds] ******************************************************",
                                "Thursday 26 March 2026  05:58:11 +0000 (0:00:01.165)       0:55:34.751 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Upgrade standbys ceph mdss cluster] **************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  05:58:15 +0000 (0:00:04.539)       0:55:39.291 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  05:58:17 +0000 (0:00:01.442)       0:55:40.734 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  05:58:18 +0000 (0:00:01.673)       0:55:42.407 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  05:58:20 +0000 (0:00:01.527)       0:55:43.936 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  05:58:21 +0000 (0:00:01.530)       0:55:45.467 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  05:58:23 +0000 (0:00:01.410)       0:55:46.878 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  05:58:24 +0000 (0:00:01.338)       0:55:48.217 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  05:58:25 +0000 (0:00:01.342)       0:55:49.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  05:58:27 +0000 (0:00:01.272)       0:55:50.832 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  05:58:28 +0000 (0:00:01.355)       0:55:52.187 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  05:58:30 +0000 (0:00:01.742)       0:55:53.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  05:58:31 +0000 (0:00:01.360)       0:55:55.291 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  05:58:34 +0000 (0:00:03.016)       0:55:58.308 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  05:58:36 +0000 (0:00:01.430)       0:55:59.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  05:58:37 +0000 (0:00:01.644)       0:56:01.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  05:58:38 +0000 (0:00:01.165)       0:56:02.548 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 05:58:32.222055', 'end': '2026-03-26 05:58:32.279175', 'delta': '0:00:00.057120', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 05:58:32.799939', 'end': '2026-03-26 05:58:32.846328', 'delta': '0:00:00.046389', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 05:58:33.349459', 'end': '2026-03-26 05:58:33.398492', 'delta': '0:00:00.049033', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  05:58:40 +0000 (0:00:01.255)       0:56:03.803 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  05:58:41 +0000 (0:00:01.450)       0:56:05.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  05:58:42 +0000 (0:00:01.250)       0:56:06.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  05:58:44 +0000 (0:00:01.373)       0:56:07.878 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:58:46 +0000 (0:00:02.318)       0:56:10.196 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  05:58:47 +0000 (0:00:01.273)       0:56:11.470 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  05:58:48 +0000 (0:00:01.155)       0:56:12.626 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  05:58:50 +0000 (0:00:01.222)       0:56:13.848 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  05:58:51 +0000 (0:00:01.240)       0:56:15.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  05:58:52 +0000 (0:00:01.242)       0:56:16.331 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  05:58:53 +0000 (0:00:01.320)       0:56:17.652 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  05:58:55 +0000 (0:00:01.617)       0:56:19.269 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  05:58:56 +0000 (0:00:01.279)       0:56:20.549 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  05:58:58 +0000 (0:00:01.264)       0:56:21.814 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  05:58:59 +0000 (0:00:01.314)       0:56:23.128 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  05:59:00 +0000 (0:00:01.509)       0:56:24.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  05:59:02 +0000 (0:00:01.559)       0:56:26.197 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  05:59:04 +0000 (0:00:01.668)       0:56:27.866 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:59:05 +0000 (0:00:01.240)       0:56:29.107 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:59:07 +0000 (0:00:01.618)       0:56:30.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  05:59:08 +0000 (0:00:01.246)       0:56:31.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  05:59:09 +0000 (0:00:01.364)       0:56:33.336 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  05:59:10 +0000 (0:00:01.295)       0:56:34.632 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  05:59:13 +0000 (0:00:02.267)       0:56:36.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  05:59:14 +0000 (0:00:01.597)       0:56:38.497 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  05:59:16 +0000 (0:00:01.323)       0:56:39.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  05:59:17 +0000 (0:00:01.335)       0:56:41.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  05:59:18 +0000 (0:00:01.262)       0:56:42.419 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  05:59:19 +0000 (0:00:01.224)       0:56:43.644 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  05:59:21 +0000 (0:00:01.347)       0:56:44.991 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  05:59:23 +0000 (0:00:01.796)       0:56:46.788 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  05:59:24 +0000 (0:00:01.420)       0:56:48.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  05:59:25 +0000 (0:00:01.421)       0:56:49.630 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  05:59:27 +0000 (0:00:01.300)       0:56:50.931 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  05:59:29 +0000 (0:00:01.936)       0:56:52.868 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  05:59:31 +0000 (0:00:02.203)       0:56:55.071 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Prevent restarts from the packaging] *************************************",
                                "Thursday 26 March 2026  05:59:34 +0000 (0:00:02.707)       0:56:57.778 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  05:59:35 +0000 (0:00:01.214)       0:56:58.993 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  05:59:36 +0000 (0:00:01.540)       0:57:00.533 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  05:59:38 +0000 (0:00:01.267)       0:57:01.800 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  05:59:39 +0000 (0:00:01.300)       0:57:03.101 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  05:59:41 +0000 (0:00:01.581)       0:57:04.682 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  05:59:42 +0000 (0:00:01.705)       0:57:06.387 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  05:59:44 +0000 (0:00:01.631)       0:57:08.019 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  05:59:45 +0000 (0:00:01.243)       0:57:09.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  05:59:46 +0000 (0:00:01.247)       0:57:10.510 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  05:59:48 +0000 (0:00:01.277)       0:57:11.787 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  05:59:49 +0000 (0:00:01.612)       0:57:13.400 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  05:59:51 +0000 (0:00:01.638)       0:57:15.039 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  05:59:53 +0000 (0:00:01.646)       0:57:16.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  05:59:54 +0000 (0:00:01.256)       0:57:17.942 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  05:59:55 +0000 (0:00:01.324)       0:57:19.266 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  05:59:56 +0000 (0:00:01.264)       0:57:20.531 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  05:59:58 +0000 (0:00:01.272)       0:57:21.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  05:59:59 +0000 (0:00:01.682)       0:57:23.486 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  06:00:01 +0000 (0:00:01.238)       0:57:24.725 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  06:00:02 +0000 (0:00:01.668)       0:57:26.395 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  06:00:03 +0000 (0:00:01.243)       0:57:27.638 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  06:00:05 +0000 (0:00:01.346)       0:57:28.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  06:00:06 +0000 (0:00:01.278)       0:57:30.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  06:00:07 +0000 (0:00:01.298)       0:57:31.562 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  06:00:09 +0000 (0:00:01.164)       0:57:32.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  06:00:10 +0000 (0:00:01.284)       0:57:34.011 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  06:00:11 +0000 (0:00:01.170)       0:57:35.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  06:00:12 +0000 (0:00:01.203)       0:57:36.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  06:00:13 +0000 (0:00:01.167)       0:57:37.552 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  06:00:15 +0000 (0:00:01.195)       0:57:38.748 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  06:00:16 +0000 (0:00:01.229)       0:57:39.978 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  06:00:17 +0000 (0:00:01.673)       0:57:41.651 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  06:00:19 +0000 (0:00:01.238)       0:57:42.890 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  06:00:20 +0000 (0:00:01.227)       0:57:44.118 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  06:00:22 +0000 (0:00:02.049)       0:57:46.167 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  06:00:24 +0000 (0:00:02.379)       0:57:48.546 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  06:00:26 +0000 (0:00:01.473)       0:57:50.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  06:00:27 +0000 (0:00:01.306)       0:57:51.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  06:00:28 +0000 (0:00:01.248)       0:57:52.575 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  06:00:30 +0000 (0:00:01.880)       0:57:54.455 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  06:00:32 +0000 (0:00:01.591)       0:57:56.047 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  06:00:33 +0000 (0:00:01.242)       0:57:57.289 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  06:00:34 +0000 (0:00:01.351)       0:57:58.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  06:00:36 +0000 (0:00:01.187)       0:57:59.829 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  06:00:37 +0000 (0:00:01.249)       0:58:01.078 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  06:00:39 +0000 (0:00:01.920)       0:58:02.999 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  06:00:40 +0000 (0:00:01.223)       0:58:04.222 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  06:00:41 +0000 (0:00:01.258)       0:58:05.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  06:00:43 +0000 (0:00:01.271)       0:58:06.752 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  06:00:44 +0000 (0:00:01.239)       0:58:07.992 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  06:00:45 +0000 (0:00:01.323)       0:58:09.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  06:00:46 +0000 (0:00:01.289)       0:58:10.605 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  06:00:49 +0000 (0:00:02.705)       0:58:13.310 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  06:00:50 +0000 (0:00:01.215)       0:58:14.526 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  06:00:52 +0000 (0:00:01.462)       0:58:15.988 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  06:00:53 +0000 (0:00:01.268)       0:58:17.257 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  06:00:54 +0000 (0:00:01.286)       0:58:18.543 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  06:00:56 +0000 (0:00:01.272)       0:58:19.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  06:00:57 +0000 (0:00:01.256)       0:58:21.072 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  06:00:58 +0000 (0:00:01.263)       0:58:22.336 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  06:01:00 +0000 (0:00:01.352)       0:58:23.688 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  06:01:01 +0000 (0:00:01.283)       0:58:24.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  06:01:02 +0000 (0:00:01.310)       0:58:26.283 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  06:01:03 +0000 (0:00:01.254)       0:58:27.537 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  06:01:05 +0000 (0:00:01.226)       0:58:28.763 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  06:01:11 +0000 (0:00:06.635)       0:58:35.399 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  06:01:13 +0000 (0:00:01.343)       0:58:36.742 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  06:01:14 +0000 (0:00:01.652)       0:58:38.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  06:01:16 +0000 (0:00:02.002)       0:58:40.397 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  06:01:18 +0000 (0:00:01.311)       0:58:41.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  06:01:19 +0000 (0:00:01.260)       0:58:42.969 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  06:01:20 +0000 (0:00:01.598)       0:58:44.567 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  06:01:22 +0000 (0:00:01.212)       0:58:45.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  06:01:23 +0000 (0:00:01.208)       0:58:46.987 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  06:01:24 +0000 (0:00:01.235)       0:58:48.222 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  06:01:25 +0000 (0:00:01.290)       0:58:49.513 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  06:01:27 +0000 (0:00:01.316)       0:58:50.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  06:01:28 +0000 (0:00:01.315)       0:58:52.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  06:01:29 +0000 (0:00:01.304)       0:58:53.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  06:01:31 +0000 (0:00:01.304)       0:58:54.755 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  06:01:35 +0000 (0:00:04.640)       0:58:59.396 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  06:01:37 +0000 (0:00:01.357)       0:59:00.753 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  06:01:42 +0000 (0:00:05.214)       0:59:05.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  06:01:43 +0000 (0:00:01.242)       0:59:07.211 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:01:45 +0000 (0:00:01.606)       0:59:08.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:01:46 +0000 (0:00:01.271)       0:59:10.089 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:01:47 +0000 (0:00:01.305)       0:59:11.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:01:49 +0000 (0:00:01.288)       0:59:12.684 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:01:50 +0000 (0:00:01.391)       0:59:14.075 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:01:51 +0000 (0:00:01.420)       0:59:15.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:01:53 +0000 (0:00:01.426)       0:59:16.922 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:01:55 +0000 (0:00:01.794)       0:59:18.717 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:01:56 +0000 (0:00:01.338)       0:59:20.055 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  06:01:57 +0000 (0:00:01.506)       0:59:21.562 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Thursday 26 March 2026  06:01:59 +0000 (0:00:01.889)       0:59:23.451 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  06:02:01 +0000 (0:00:01.232)       0:59:24.684 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Thursday 26 March 2026  06:02:02 +0000 (0:00:01.436)       0:59:26.121 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:02:04 +0000 (0:00:01.980)       0:59:28.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:02:07 +0000 (0:00:03.222)       0:59:31.325 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Thursday 26 March 2026  06:02:09 +0000 (0:00:02.087)       0:59:33.413 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Thursday 26 March 2026  06:02:11 +0000 (0:00:01.583)       0:59:34.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Thursday 26 March 2026  06:02:12 +0000 (0:00:01.227)       0:59:36.224 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Thursday 26 March 2026  06:02:14 +0000 (0:00:01.570)       0:59:37.794 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  06:02:15 +0000 (0:00:01.442)       0:59:39.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Thursday 26 March 2026  06:02:17 +0000 (0:00:02.253)       0:59:41.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Thursday 26 March 2026  06:02:19 +0000 (0:00:02.075)       0:59:43.567 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Thursday 26 March 2026  06:02:22 +0000 (0:00:02.371)       0:59:45.939 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Thursday 26 March 2026  06:02:25 +0000 (0:00:03.416)       0:59:49.355 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set max_mds] *************************************************************",
                                "Thursday 26 March 2026  06:02:27 +0000 (0:00:01.849)       0:59:51.205 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph rgws cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  06:02:31 +0000 (0:00:03.502)       0:59:54.708 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  06:02:32 +0000 (0:00:01.142)       0:59:55.850 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  06:02:33 +0000 (0:00:01.452)       0:59:57.302 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:02:34 +0000 (0:00:01.123)       0:59:58.426 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:02:36 +0000 (0:00:01.431)       0:59:59.857 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  06:02:37 +0000 (0:00:01.211)       1:00:01.069 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  06:02:38 +0000 (0:00:01.122)       1:00:02.192 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  06:02:39 +0000 (0:00:01.165)       1:00:03.358 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  06:02:40 +0000 (0:00:01.145)       1:00:04.503 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  06:02:42 +0000 (0:00:01.245)       1:00:05.749 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  06:02:43 +0000 (0:00:01.699)       1:00:07.448 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  06:02:45 +0000 (0:00:01.259)       1:00:08.707 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  06:02:47 +0000 (0:00:02.888)       1:00:11.596 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  06:02:49 +0000 (0:00:01.503)       1:00:13.099 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  06:02:51 +0000 (0:00:01.923)       1:00:15.022 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  06:02:52 +0000 (0:00:01.212)       1:00:16.235 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:02:45.573480', 'end': '2026-03-26 06:02:45.621529', 'delta': '0:00:00.048049', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:02:46.135117', 'end': '2026-03-26 06:02:46.183458', 'delta': '0:00:00.048341', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:02:46.702167', 'end': '2026-03-26 06:02:46.750702', 'delta': '0:00:00.048535', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  06:02:53 +0000 (0:00:01.174)       1:00:17.410 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  06:02:55 +0000 (0:00:01.248)       1:00:18.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  06:02:56 +0000 (0:00:01.697)       1:00:20.357 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  06:02:57 +0000 (0:00:01.188)       1:00:21.546 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:02:59 +0000 (0:00:01.989)       1:00:23.535 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  06:03:01 +0000 (0:00:01.183)       1:00:24.718 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  06:03:02 +0000 (0:00:01.161)       1:00:25.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:03:03 +0000 (0:00:01.210)       1:00:27.090 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  06:03:04 +0000 (0:00:01.113)       1:00:28.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  06:03:05 +0000 (0:00:01.133)       1:00:29.338 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  06:03:06 +0000 (0:00:01.187)       1:00:30.526 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  06:03:07 +0000 (0:00:01.125)       1:00:31.652 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  06:03:09 +0000 (0:00:01.210)       1:00:32.863 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  06:03:10 +0000 (0:00:01.190)       1:00:34.053 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  06:03:11 +0000 (0:00:01.166)       1:00:35.219 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  06:03:13 +0000 (0:00:01.541)       1:00:36.761 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082', 'dm-uuid-LVM-8hKVl461SF70Ai5uMDmNdT5BP20Vvkg8AxHs2aTbdloCZd5zRhurro2iqvFnFzRY'], 'uuids': ['c579629d-afc9-41d5-a76c-63e3abbafb40'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8', 'scsi-SQEMU_QEMU_HARDDISK_2dae49df-17cb-48b5-9940-ec5e7ec792d8'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '2dae49df', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-2XKfyD-kvYx-XaUk-IA1D-OFMu-auWL-FeQHCw', 'scsi-0QEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80', 'scsi-SQEMU_QEMU_HARDDISK_d11e4e4a-db1d-44df-8da9-5de7e993dd80'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-13-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS', 'dm-uuid-CRYPT-LUKS2-aef43475035b4229a7d71e3432ab4dcb-y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a-osd--block--93e8c9a2--b6ff--5fe0--a79e--2922336c3e0a', 'dm-uuid-LVM-NfuOn4R5AkCZoZBaGfCwjgSejX4qlSlby5xuVgNQ7T0MWashc4xC7nHJ3VUNBCRS'], 'uuids': ['aef43475-035b-4229-a7d7-1e3432ab4dcb'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'd11e4e4a', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['y5xuVg-NQ7T-0MWa-shc4-xC7n-HJ3V-UNBCRS']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-dxNnp3-HdCF-97hz-w17k-bHEu-opcA-g4y34j', 'scsi-0QEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331', 'scsi-SQEMU_QEMU_HARDDISK_863ba5d2-7e2f-4393-95a6-83543745d331'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '863ba5d2', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--e2623153--bc41--510f--8884--ef957bb96082-osd--block--e2623153--bc41--510f--8884--ef957bb96082']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'ce600cf2', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1', 'scsi-SQEMU_QEMU_HARDDISK_ce600cf2-62c4-44aa-8248-5535335c6519-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY', 'dm-uuid-CRYPT-LUKS2-c579629dafc941d5a76c63e3abbafb40-AxHs2a-Tbdl-oCZd-5zRh-urro-2iqv-FnFzRY'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  06:03:14 +0000 (0:00:01.432)       1:00:38.193 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  06:03:16 +0000 (0:00:01.541)       1:00:39.734 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:03:17 +0000 (0:00:01.241)       1:00:40.976 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:03:18 +0000 (0:00:01.543)       1:00:42.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:03:20 +0000 (0:00:01.218)       1:00:43.739 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:03:21 +0000 (0:00:01.265)       1:00:45.005 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  06:03:22 +0000 (0:00:01.155)       1:00:46.160 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  06:03:24 +0000 (0:00:02.024)       1:00:48.185 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  06:03:25 +0000 (0:00:01.193)       1:00:49.379 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:03:26 +0000 (0:00:01.136)       1:00:50.516 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:03:27 +0000 (0:00:01.122)       1:00:51.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:03:29 +0000 (0:00:01.242)       1:00:52.881 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:03:30 +0000 (0:00:01.155)       1:00:54.037 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:03:31 +0000 (0:00:01.210)       1:00:55.247 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:03:33 +0000 (0:00:01.437)       1:00:56.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:03:34 +0000 (0:00:01.470)       1:00:58.155 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:03:35 +0000 (0:00:01.422)       1:00:59.578 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:03:37 +0000 (0:00:01.176)       1:01:00.755 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  06:03:38 +0000 (0:00:01.330)       1:01:02.085 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  06:03:40 +0000 (0:00:02.193)       1:01:04.278 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw when upgrading from stable-3.2] ****************************",
                                "Thursday 26 March 2026  06:03:43 +0000 (0:00:02.646)       1:01:06.924 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 1)] ***************************************************",
                                "Thursday 26 March 2026  06:03:45 +0000 (0:00:02.315)       1:01:09.240 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 2)] ***************************************************",
                                "Thursday 26 March 2026  06:03:48 +0000 (0:00:02.873)       1:01:12.113 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  06:03:50 +0000 (0:00:02.383)       1:01:14.497 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  06:03:51 +0000 (0:00:01.109)       1:01:15.606 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  06:03:53 +0000 (0:00:01.146)       1:01:16.753 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  06:03:54 +0000 (0:00:01.121)       1:01:17.875 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  06:03:55 +0000 (0:00:01.497)       1:01:19.372 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  06:03:57 +0000 (0:00:01.557)       1:01:20.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  06:03:58 +0000 (0:00:01.501)       1:01:22.431 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  06:03:59 +0000 (0:00:01.125)       1:01:23.556 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  06:04:01 +0000 (0:00:01.125)       1:01:24.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  06:04:02 +0000 (0:00:01.108)       1:01:25.790 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  06:04:03 +0000 (0:00:01.514)       1:01:27.305 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  06:04:05 +0000 (0:00:01.503)       1:01:28.809 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  06:04:06 +0000 (0:00:01.183)       1:01:29.993 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  06:04:07 +0000 (0:00:01.100)       1:01:31.093 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  06:04:08 +0000 (0:00:01.127)       1:01:32.220 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  06:04:09 +0000 (0:00:01.126)       1:01:33.347 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  06:04:10 +0000 (0:00:01.224)       1:01:34.571 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  06:04:12 +0000 (0:00:01.119)       1:01:35.691 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  06:04:13 +0000 (0:00:01.142)       1:01:36.833 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  06:04:14 +0000 (0:00:01.139)       1:01:37.973 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  06:04:15 +0000 (0:00:01.178)       1:01:39.151 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  06:04:16 +0000 (0:00:01.186)       1:01:40.338 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  06:04:17 +0000 (0:00:01.148)       1:01:41.487 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  06:04:18 +0000 (0:00:01.114)       1:01:42.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  06:04:20 +0000 (0:00:01.242)       1:01:43.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  06:04:21 +0000 (0:00:01.216)       1:01:45.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  06:04:22 +0000 (0:00:01.141)       1:01:46.202 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  06:04:23 +0000 (0:00:01.129)       1:01:47.331 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  06:04:24 +0000 (0:00:01.112)       1:01:48.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  06:04:25 +0000 (0:00:01.131)       1:01:49.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  06:04:27 +0000 (0:00:01.103)       1:01:50.679 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  06:04:28 +0000 (0:00:01.125)       1:01:51.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  06:04:29 +0000 (0:00:01.126)       1:01:52.931 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  06:04:30 +0000 (0:00:01.168)       1:01:54.100 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  06:04:32 +0000 (0:00:01.900)       1:01:56.000 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  06:04:34 +0000 (0:00:02.277)       1:01:58.278 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  06:04:35 +0000 (0:00:01.243)       1:01:59.521 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  06:04:36 +0000 (0:00:01.132)       1:02:00.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  06:04:38 +0000 (0:00:01.128)       1:02:01.783 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  06:04:39 +0000 (0:00:01.787)       1:02:03.571 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  06:04:41 +0000 (0:00:01.497)       1:02:05.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  06:04:42 +0000 (0:00:01.151)       1:02:06.219 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  06:04:43 +0000 (0:00:01.131)       1:02:07.351 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  06:04:44 +0000 (0:00:01.149)       1:02:08.500 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  06:04:45 +0000 (0:00:01.145)       1:02:09.645 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  06:04:47 +0000 (0:00:01.938)       1:02:11.584 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  06:04:49 +0000 (0:00:01.231)       1:02:12.816 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  06:04:50 +0000 (0:00:01.358)       1:02:14.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  06:04:51 +0000 (0:00:01.202)       1:02:15.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  06:04:52 +0000 (0:00:01.198)       1:02:16.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  06:04:54 +0000 (0:00:01.230)       1:02:17.807 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  06:04:55 +0000 (0:00:01.201)       1:02:19.008 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  06:04:57 +0000 (0:00:02.504)       1:02:21.513 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  06:04:59 +0000 (0:00:01.152)       1:02:22.665 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  06:05:00 +0000 (0:00:01.113)       1:02:23.779 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  06:05:01 +0000 (0:00:01.182)       1:02:24.962 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  06:05:02 +0000 (0:00:01.161)       1:02:26.124 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  06:05:03 +0000 (0:00:01.185)       1:02:27.310 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  06:05:04 +0000 (0:00:01.160)       1:02:28.470 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  06:05:05 +0000 (0:00:01.153)       1:02:29.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  06:05:07 +0000 (0:00:01.262)       1:02:30.886 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  06:05:08 +0000 (0:00:01.157)       1:02:32.043 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  06:05:09 +0000 (0:00:01.163)       1:02:33.207 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  06:05:10 +0000 (0:00:01.177)       1:02:34.385 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  06:05:11 +0000 (0:00:01.115)       1:02:35.500 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  06:05:18 +0000 (0:00:06.698)       1:02:42.199 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  06:05:19 +0000 (0:00:01.155)       1:02:43.354 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  06:05:21 +0000 (0:00:01.521)       1:02:44.875 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  06:05:23 +0000 (0:00:01.949)       1:02:46.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  06:05:24 +0000 (0:00:01.158)       1:02:47.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  06:05:25 +0000 (0:00:01.278)       1:02:49.261 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  06:05:26 +0000 (0:00:01.123)       1:02:50.385 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  06:05:27 +0000 (0:00:01.139)       1:02:51.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  06:05:28 +0000 (0:00:01.131)       1:02:52.657 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  06:05:30 +0000 (0:00:01.101)       1:02:53.758 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  06:05:31 +0000 (0:00:01.206)       1:02:54.965 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  06:05:32 +0000 (0:00:01.150)       1:02:56.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  06:05:33 +0000 (0:00:01.120)       1:02:57.236 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  06:05:34 +0000 (0:00:01.133)       1:02:58.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  06:05:35 +0000 (0:00:01.172)       1:02:59.542 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  06:05:40 +0000 (0:00:04.512)       1:03:04.055 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  06:05:41 +0000 (0:00:01.148)       1:03:05.204 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  06:05:46 +0000 (0:00:05.259)       1:03:10.463 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  06:05:48 +0000 (0:00:01.309)       1:03:11.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:05:49 +0000 (0:00:01.196)       1:03:12.970 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:05:50 +0000 (0:00:01.248)       1:03:14.218 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:05:51 +0000 (0:00:01.127)       1:03:15.346 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:05:52 +0000 (0:00:01.210)       1:03:16.557 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:05:54 +0000 (0:00:01.257)       1:03:17.814 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:05:55 +0000 (0:00:01.473)       1:03:19.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:05:57 +0000 (0:00:01.405)       1:03:20.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:05:58 +0000 (0:00:01.410)       1:03:22.104 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:05:59 +0000 (0:00:01.170)       1:03:23.275 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  06:06:00 +0000 (0:00:01.369)       1:03:24.645 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  06:06:02 +0000 (0:00:01.946)       1:03:26.591 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:06:04 +0000 (0:00:01.472)       1:03:28.064 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:06:07 +0000 (0:00:03.236)       1:03:31.300 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Thursday 26 March 2026  06:06:09 +0000 (0:00:02.013)       1:03:33.314 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Thursday 26 March 2026  06:06:10 +0000 (0:00:01.169)       1:03:34.484 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Thursday 26 March 2026  06:06:12 +0000 (0:00:01.484)       1:03:35.968 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Thursday 26 March 2026  06:06:13 +0000 (0:00:01.599)       1:03:37.568 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:06:19 +0000 (0:00:05.222)       1:03:42.790 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:06:22 +0000 (0:00:03.171)       1:03:45.961 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Thursday 26 March 2026  06:06:24 +0000 (0:00:02.040)       1:03:48.002 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Thursday 26 March 2026  06:06:25 +0000 (0:00:01.630)       1:03:49.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Thursday 26 March 2026  06:06:27 +0000 (0:00:01.612)       1:03:51.245 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Thursday 26 March 2026  06:06:29 +0000 (0:00:01.666)       1:03:52.912 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Thursday 26 March 2026  06:07:00 +0000 (0:00:31.259)       1:04:24.172 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Thursday 26 March 2026  06:07:01 +0000 (0:00:01.146)       1:04:25.319 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Thursday 26 March 2026  06:07:02 +0000 (0:00:01.146)       1:04:26.465 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Thursday 26 March 2026  06:07:04 +0000 (0:00:01.491)       1:04:27.957 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  06:07:05 +0000 (0:00:01.521)       1:04:29.478 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Thursday 26 March 2026  06:07:07 +0000 (0:00:02.099)       1:04:31.577 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Thursday 26 March 2026  06:07:09 +0000 (0:00:02.027)       1:04:33.605 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Thursday 26 March 2026  06:07:12 +0000 (0:00:02.309)       1:04:35.914 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph rgws cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  06:07:15 +0000 (0:00:03.155)       1:04:39.070 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  06:07:16 +0000 (0:00:01.119)       1:04:40.189 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  06:07:17 +0000 (0:00:01.429)       1:04:41.619 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:07:19 +0000 (0:00:01.150)       1:04:42.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:07:20 +0000 (0:00:01.465)       1:04:44.236 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  06:07:21 +0000 (0:00:01.143)       1:04:45.379 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  06:07:22 +0000 (0:00:01.125)       1:04:46.505 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  06:07:24 +0000 (0:00:01.207)       1:04:47.713 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  06:07:25 +0000 (0:00:01.143)       1:04:48.857 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  06:07:26 +0000 (0:00:01.132)       1:04:49.990 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  06:07:28 +0000 (0:00:02.144)       1:04:52.134 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  06:07:29 +0000 (0:00:01.245)       1:04:53.379 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  06:07:32 +0000 (0:00:02.918)       1:04:56.298 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  06:07:34 +0000 (0:00:01.492)       1:04:57.791 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  06:07:35 +0000 (0:00:01.734)       1:04:59.526 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  06:07:37 +0000 (0:00:01.227)       1:05:00.753 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:07:30.238507', 'end': '2026-03-26 06:07:30.291933', 'delta': '0:00:00.053426', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:07:30.798086', 'end': '2026-03-26 06:07:30.854639', 'delta': '0:00:00.056553', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:07:31.399026', 'end': '2026-03-26 06:07:31.431938', 'delta': '0:00:00.032912', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  06:07:38 +0000 (0:00:01.202)       1:05:01.955 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  06:07:39 +0000 (0:00:01.272)       1:05:03.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  06:07:40 +0000 (0:00:01.269)       1:05:04.497 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  06:07:41 +0000 (0:00:01.151)       1:05:05.648 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:07:43 +0000 (0:00:01.996)       1:05:07.645 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  06:07:45 +0000 (0:00:01.162)       1:05:08.808 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  06:07:46 +0000 (0:00:01.121)       1:05:09.929 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:07:48 +0000 (0:00:01.740)       1:05:11.669 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  06:07:49 +0000 (0:00:01.177)       1:05:12.847 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  06:07:50 +0000 (0:00:01.128)       1:05:13.975 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  06:07:51 +0000 (0:00:01.151)       1:05:15.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  06:07:52 +0000 (0:00:01.147)       1:05:16.274 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  06:07:53 +0000 (0:00:01.172)       1:05:17.447 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  06:07:55 +0000 (0:00:01.213)       1:05:18.661 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  06:07:56 +0000 (0:00:01.163)       1:05:19.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  06:07:57 +0000 (0:00:01.331)       1:05:21.155 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543', 'dm-uuid-LVM-O1aEkSX5V2TgXKGnqX2peNd9dQhi04NAZJyEqlgfRLjtJKN8JwRgDI1ZPO4R3wgt'], 'uuids': ['1d39f6c5-1f6c-4630-99cd-a410ca5e45d8'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44', 'scsi-SQEMU_QEMU_HARDDISK_7e352b46-e023-45cf-8a88-51cc46240a44'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7e352b46', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-eoBjP8-dDdJ-3FQm-pH7P-5B72-c1L3-mABWfX', 'scsi-0QEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab', 'scsi-SQEMU_QEMU_HARDDISK_7db5f133-fe7b-42a4-ad57-b076dc1856ab'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-09-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG', 'dm-uuid-CRYPT-LUKS2-741ece0a80b8415aa2e2dcc695db5f53-2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--a652979e--9f40--503a--bbc8--6de5e605991e-osd--block--a652979e--9f40--503a--bbc8--6de5e605991e', 'dm-uuid-LVM-86WEu6duX2Pejl3asW6viK3fsh4aqvqg2h2U7SLeR6PGwru1xY81U9rrCs8siESG'], 'uuids': ['741ece0a-80b8-415a-a2e2-dcc695db5f53'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '7db5f133', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['2h2U7S-LeR6-PGwr-u1xY-81U9-rrCs-8siESG']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-Oy69b4-OcVV-F2KD-vi5G-C8ns-n3Cu-1PhYTB', 'scsi-0QEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263', 'scsi-SQEMU_QEMU_HARDDISK_a52ec37c-b4ea-4f83-9b16-3c0f6ce85263'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': 'a52ec37c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--b5eee7c3--8883--5bbe--be5a--75726e822543-osd--block--b5eee7c3--8883--5bbe--be5a--75726e822543']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '48d73a84', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1', 'scsi-SQEMU_QEMU_HARDDISK_48d73a84-835d-480a-92c3-3edf7ed142ea-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt', 'dm-uuid-CRYPT-LUKS2-1d39f6c51f6c463099cda410ca5e45d8-ZJyEql-gfRL-jtJK-N8Jw-RgDI-1ZPO-4R3wgt'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  06:07:58 +0000 (0:00:01.444)       1:05:22.600 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  06:08:00 +0000 (0:00:01.517)       1:05:24.119 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:08:01 +0000 (0:00:01.150)       1:05:25.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:08:03 +0000 (0:00:01.472)       1:05:26.743 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:08:04 +0000 (0:00:01.310)       1:05:28.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:08:05 +0000 (0:00:01.315)       1:05:29.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  06:08:06 +0000 (0:00:01.125)       1:05:30.495 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  06:08:08 +0000 (0:00:01.668)       1:05:32.164 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  06:08:10 +0000 (0:00:01.664)       1:05:33.829 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:08:11 +0000 (0:00:01.166)       1:05:34.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:08:12 +0000 (0:00:01.143)       1:05:36.139 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:08:13 +0000 (0:00:01.134)       1:05:37.274 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:08:14 +0000 (0:00:01.136)       1:05:38.410 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:08:16 +0000 (0:00:01.297)       1:05:39.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:08:17 +0000 (0:00:01.393)       1:05:41.101 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:08:18 +0000 (0:00:01.517)       1:05:42.619 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:08:20 +0000 (0:00:01.362)       1:05:43.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:08:21 +0000 (0:00:00.921)       1:05:44.903 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  06:08:22 +0000 (0:00:01.229)       1:05:46.133 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  06:08:24 +0000 (0:00:01.797)       1:05:47.930 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw when upgrading from stable-3.2] ****************************",
                                "Thursday 26 March 2026  06:08:26 +0000 (0:00:02.534)       1:05:50.464 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 1)] ***************************************************",
                                "Thursday 26 March 2026  06:08:28 +0000 (0:00:01.943)       1:05:52.408 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 2)] ***************************************************",
                                "Thursday 26 March 2026  06:08:32 +0000 (0:00:03.633)       1:05:56.041 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  06:08:34 +0000 (0:00:01.931)       1:05:57.973 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  06:08:35 +0000 (0:00:01.119)       1:05:59.093 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  06:08:36 +0000 (0:00:01.099)       1:06:00.192 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  06:08:37 +0000 (0:00:01.128)       1:06:01.321 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  06:08:39 +0000 (0:00:01.588)       1:06:02.909 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  06:08:40 +0000 (0:00:01.522)       1:06:04.431 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  06:08:42 +0000 (0:00:01.521)       1:06:05.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  06:08:43 +0000 (0:00:01.161)       1:06:07.114 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  06:08:44 +0000 (0:00:01.145)       1:06:08.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  06:08:45 +0000 (0:00:01.167)       1:06:09.427 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  06:08:47 +0000 (0:00:01.557)       1:06:10.985 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  06:08:48 +0000 (0:00:01.513)       1:06:12.499 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  06:08:49 +0000 (0:00:00.778)       1:06:13.277 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  06:08:50 +0000 (0:00:00.806)       1:06:14.083 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  06:08:51 +0000 (0:00:00.838)       1:06:14.922 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  06:08:52 +0000 (0:00:00.832)       1:06:15.754 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  06:08:52 +0000 (0:00:00.845)       1:06:16.600 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  06:08:53 +0000 (0:00:00.846)       1:06:17.446 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  06:08:54 +0000 (0:00:00.815)       1:06:18.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  06:08:55 +0000 (0:00:00.773)       1:06:19.035 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  06:08:56 +0000 (0:00:00.814)       1:06:19.850 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  06:08:56 +0000 (0:00:00.805)       1:06:20.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  06:08:57 +0000 (0:00:00.758)       1:06:21.415 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  06:08:58 +0000 (0:00:00.853)       1:06:22.268 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  06:08:59 +0000 (0:00:00.786)       1:06:23.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  06:09:00 +0000 (0:00:00.765)       1:06:23.820 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  06:09:00 +0000 (0:00:00.764)       1:06:24.585 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  06:09:01 +0000 (0:00:00.809)       1:06:25.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  06:09:02 +0000 (0:00:00.805)       1:06:26.200 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  06:09:03 +0000 (0:00:00.864)       1:06:27.065 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  06:09:04 +0000 (0:00:00.840)       1:06:27.905 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  06:09:05 +0000 (0:00:00.811)       1:06:28.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  06:09:05 +0000 (0:00:00.803)       1:06:29.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  06:09:06 +0000 (0:00:00.774)       1:06:30.295 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  06:09:08 +0000 (0:00:01.596)       1:06:31.892 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  06:09:10 +0000 (0:00:01.890)       1:06:33.782 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  06:09:11 +0000 (0:00:01.174)       1:06:34.956 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  06:09:12 +0000 (0:00:01.164)       1:06:36.121 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  06:09:13 +0000 (0:00:01.144)       1:06:37.265 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  06:09:15 +0000 (0:00:01.872)       1:06:39.138 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  06:09:16 +0000 (0:00:01.439)       1:06:40.578 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  06:09:18 +0000 (0:00:01.715)       1:06:42.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  06:09:19 +0000 (0:00:00.823)       1:06:43.117 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  06:09:20 +0000 (0:00:00.793)       1:06:43.910 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  06:09:21 +0000 (0:00:01.138)       1:06:45.049 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  06:09:23 +0000 (0:00:01.847)       1:06:46.897 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  06:09:24 +0000 (0:00:01.138)       1:06:48.035 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  06:09:25 +0000 (0:00:01.135)       1:06:49.171 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  06:09:26 +0000 (0:00:01.141)       1:06:50.313 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  06:09:27 +0000 (0:00:01.187)       1:06:51.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  06:09:29 +0000 (0:00:01.179)       1:06:52.681 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  06:09:29 +0000 (0:00:00.807)       1:06:53.488 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  06:09:31 +0000 (0:00:02.130)       1:06:55.619 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  06:09:32 +0000 (0:00:00.856)       1:06:56.475 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  06:09:33 +0000 (0:00:01.162)       1:06:57.637 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  06:09:35 +0000 (0:00:01.131)       1:06:58.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  06:09:36 +0000 (0:00:01.155)       1:06:59.924 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  06:09:37 +0000 (0:00:01.174)       1:07:01.099 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  06:09:38 +0000 (0:00:01.181)       1:07:02.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  06:09:39 +0000 (0:00:01.119)       1:07:03.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  06:09:40 +0000 (0:00:01.223)       1:07:04.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  06:09:42 +0000 (0:00:01.161)       1:07:05.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  06:09:43 +0000 (0:00:01.113)       1:07:06.900 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  06:09:44 +0000 (0:00:00.829)       1:07:07.730 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  06:09:45 +0000 (0:00:01.306)       1:07:09.036 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  06:09:51 +0000 (0:00:06.102)       1:07:15.139 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  06:09:52 +0000 (0:00:01.117)       1:07:16.256 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  06:09:54 +0000 (0:00:01.529)       1:07:17.786 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  06:09:55 +0000 (0:00:01.707)       1:07:19.494 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  06:09:56 +0000 (0:00:00.819)       1:07:20.314 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  06:09:57 +0000 (0:00:00.766)       1:07:21.080 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  06:09:58 +0000 (0:00:00.771)       1:07:21.851 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  06:09:59 +0000 (0:00:00.806)       1:07:22.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  06:09:59 +0000 (0:00:00.803)       1:07:23.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  06:10:00 +0000 (0:00:00.810)       1:07:24.272 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  06:10:01 +0000 (0:00:00.829)       1:07:25.102 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  06:10:02 +0000 (0:00:00.815)       1:07:25.917 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  06:10:03 +0000 (0:00:00.764)       1:07:26.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  06:10:03 +0000 (0:00:00.806)       1:07:27.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  06:10:04 +0000 (0:00:00.810)       1:07:28.300 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  06:10:08 +0000 (0:00:04.119)       1:07:32.420 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  06:10:09 +0000 (0:00:00.795)       1:07:33.216 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  06:10:13 +0000 (0:00:04.429)       1:07:37.645 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  06:10:14 +0000 (0:00:00.807)       1:07:38.453 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:10:15 +0000 (0:00:00.793)       1:07:39.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:10:16 +0000 (0:00:00.817)       1:07:40.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:10:17 +0000 (0:00:00.815)       1:07:40.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:10:18 +0000 (0:00:00.821)       1:07:41.701 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:10:19 +0000 (0:00:01.058)       1:07:42.759 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:10:20 +0000 (0:00:01.076)       1:07:43.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:10:21 +0000 (0:00:01.131)       1:07:44.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:10:22 +0000 (0:00:01.087)       1:07:46.055 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:10:23 +0000 (0:00:00.843)       1:07:46.899 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  06:10:24 +0000 (0:00:01.042)       1:07:47.942 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  06:10:25 +0000 (0:00:01.435)       1:07:49.378 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:10:26 +0000 (0:00:01.141)       1:07:50.519 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:10:30 +0000 (0:00:03.219)       1:07:53.739 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Thursday 26 March 2026  06:10:32 +0000 (0:00:01.950)       1:07:55.689 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Thursday 26 March 2026  06:10:32 +0000 (0:00:00.749)       1:07:56.438 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Thursday 26 March 2026  06:10:34 +0000 (0:00:01.315)       1:07:57.754 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Thursday 26 March 2026  06:10:35 +0000 (0:00:01.590)       1:07:59.345 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:10:40 +0000 (0:00:05.213)       1:08:04.559 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:10:43 +0000 (0:00:03.044)       1:08:07.604 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Thursday 26 March 2026  06:10:45 +0000 (0:00:01.639)       1:08:09.244 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Thursday 26 March 2026  06:10:46 +0000 (0:00:01.160)       1:08:10.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Thursday 26 March 2026  06:10:48 +0000 (0:00:01.647)       1:08:12.052 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Thursday 26 March 2026  06:10:50 +0000 (0:00:01.995)       1:08:14.047 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Thursday 26 March 2026  06:11:21 +0000 (0:00:31.450)       1:08:45.498 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Thursday 26 March 2026  06:11:22 +0000 (0:00:00.762)       1:08:46.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Thursday 26 March 2026  06:11:23 +0000 (0:00:00.772)       1:08:47.032 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Thursday 26 March 2026  06:11:24 +0000 (0:00:01.262)       1:08:48.295 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  06:11:25 +0000 (0:00:01.163)       1:08:49.458 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Thursday 26 March 2026  06:11:27 +0000 (0:00:02.068)       1:08:51.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Thursday 26 March 2026  06:11:29 +0000 (0:00:01.867)       1:08:53.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Thursday 26 March 2026  06:11:31 +0000 (0:00:02.209)       1:08:55.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph rgws cluster] ***********************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  06:11:34 +0000 (0:00:02.821)       1:08:58.425 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Thursday 26 March 2026  06:11:35 +0000 (0:00:01.179)       1:08:59.604 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Thursday 26 March 2026  06:11:37 +0000 (0:00:01.538)       1:09:01.143 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:11:38 +0000 (0:00:01.152)       1:09:02.295 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:11:40 +0000 (0:00:01.490)       1:09:03.786 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Thursday 26 March 2026  06:11:41 +0000 (0:00:01.188)       1:09:04.974 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Thursday 26 March 2026  06:11:42 +0000 (0:00:01.179)       1:09:06.154 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Thursday 26 March 2026  06:11:43 +0000 (0:00:01.194)       1:09:07.348 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Thursday 26 March 2026  06:11:44 +0000 (0:00:01.137)       1:09:08.486 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Thursday 26 March 2026  06:11:45 +0000 (0:00:01.143)       1:09:09.629 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Thursday 26 March 2026  06:11:47 +0000 (0:00:01.764)       1:09:11.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Thursday 26 March 2026  06:11:49 +0000 (0:00:01.358)       1:09:12.752 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Thursday 26 March 2026  06:11:52 +0000 (0:00:03.297)       1:09:16.050 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Thursday 26 March 2026  06:11:53 +0000 (0:00:01.406)       1:09:17.456 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Thursday 26 March 2026  06:11:55 +0000 (0:00:02.092)       1:09:19.548 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Thursday 26 March 2026  06:11:57 +0000 (0:00:01.154)       1:09:20.703 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'de9c3b4c4c57', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-26 06:11:49.624637', 'end': '2026-03-26 06:11:49.669646', 'delta': '0:00:00.045009', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['de9c3b4c4c57'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'd66b87272f8e', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-26 06:11:50.174575', 'end': '2026-03-26 06:11:50.235008', 'delta': '0:00:00.060433', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['d66b87272f8e'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'changed': False, 'stdout': 'b850f8fd4697', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-26 06:11:51.182319', 'end': '2026-03-26 06:11:51.231785', 'delta': '0:00:00.049466', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['b850f8fd4697'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Thursday 26 March 2026  06:11:58 +0000 (0:00:01.315)       1:09:22.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Thursday 26 March 2026  06:11:59 +0000 (0:00:01.280)       1:09:23.299 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Thursday 26 March 2026  06:12:00 +0000 (0:00:01.259)       1:09:24.558 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Thursday 26 March 2026  06:12:02 +0000 (0:00:01.133)       1:09:25.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:12:04 +0000 (0:00:02.020)       1:09:27.712 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Thursday 26 March 2026  06:12:05 +0000 (0:00:01.150)       1:09:28.863 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Thursday 26 March 2026  06:12:06 +0000 (0:00:01.123)       1:09:29.987 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Thursday 26 March 2026  06:12:07 +0000 (0:00:01.236)       1:09:31.223 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Thursday 26 March 2026  06:12:08 +0000 (0:00:01.181)       1:09:32.405 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Thursday 26 March 2026  06:12:09 +0000 (0:00:01.121)       1:09:33.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Thursday 26 March 2026  06:12:11 +0000 (0:00:01.182)       1:09:34.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Thursday 26 March 2026  06:12:12 +0000 (0:00:01.137)       1:09:35.846 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Thursday 26 March 2026  06:12:13 +0000 (0:00:01.163)       1:09:37.009 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Thursday 26 March 2026  06:12:14 +0000 (0:00:01.302)       1:09:38.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Thursday 26 March 2026  06:12:15 +0000 (0:00:01.202)       1:09:39.514 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Thursday 26 March 2026  06:12:17 +0000 (0:00:01.407)       1:09:40.922 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771', 'dm-uuid-LVM-Q7trkX6T9bQrenPM1EuezeEWG2QB7ffx0bNZRnQ3R81VwJTdPWktYtRAGSsXVFlp'], 'uuids': ['958c3d71-9b3b-484b-8cbf-f174ba1f6fac'], 'labels': [], 'masters': ['dm-3']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2', 'scsi-SQEMU_QEMU_HARDDISK_8ddd7966-84e6-4951-8a08-7b4fb4af2bd2'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '8ddd7966', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-FriUOI-gUEr-kmP0-nYC7-MoO0-ng3W-Ej90o7', 'scsi-0QEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d', 'scsi-SQEMU_QEMU_HARDDISK_943c088c-5b56-4173-ab64-ec81e1cc816d'], 'uuids': [], 'labels': [], 'masters': ['dm-0']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'virtual': 1, 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'uuids': ['2026-03-26-01-38-15-00'], 'labels': ['config-2'], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU DVD-ROM', 'sas_address': None, 'sas_device_handle': None, 'removable': '1', 'support_discard': '0', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'mq-deadline', 'sectors': '1012', 'sectorsize': '2048', 'size': '506.00 KB', 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-2', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD', 'dm-uuid-CRYPT-LUKS2-4b88786507c84424981e8c33baf61cbe-A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-ceph--83c4def8--4703--5f7c--9549--7666ff9f2b66-osd--block--83c4def8--4703--5f7c--9549--7666ff9f2b66', 'dm-uuid-LVM-DoNgv1c108dy4eu1pvS7TOCWbuA3UXv0A6zrFIA863mhHtIp5pUFeDHxhomhuceD'], 'uuids': ['4b887865-07c8-4424-981e-8c33baf61cbe'], 'labels': [], 'masters': ['dm-2']}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'serial': '943c088c', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41934848', 'sectorsize': '512', 'size': '20.00 GB', 'host': '', 'holders': ['A6zrFI-A863-mhHt-Ip5p-UFeD-Hxho-mhuceD']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'virtual': 1, 'links': {'ids': ['lvm-pv-uuid-xgZSV6-0wfE-zGZo-XmXe-xuiN-RWM0-U4VPgB', 'scsi-0QEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102', 'scsi-SQEMU_QEMU_HARDDISK_47760649-09e9-4ed8-8303-e5ee473a8102'], 'uuids': [], 'labels': [], 'masters': ['dm-1']}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '47760649', 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '41943040', 'sectorsize': '512', 'size': '20.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': ['ceph--1fd8de68--da37--5e01--9bf2--5a04fcdcd771-osd--block--1fd8de68--da37--5e01--9bf2--5a04fcdcd771']}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'virtual': 1, 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': 'QEMU', 'model': 'QEMU HARDDISK', 'sas_address': None, 'sas_device_handle': None, 'serial': '4fa924fa', 'removable': '0', 'support_discard': '4096', 'partitions': {'sda16': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part16'], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec'], 'labels': ['BOOT'], 'masters': []}, 'start': '227328', 'sectors': '1869825', 'sectorsize': 512, 'size': '913.00 MB', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec', 'holders': []}, 'sda14': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part14'], 'uuids': [], 'labels': [], 'masters': []}, 'start': '2048', 'sectors': '8192', 'sectorsize': 512, 'size': '4.00 MB', 'uuid': None, 'holders': []}, 'sda15': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part15'], 'uuids': ['5C78-612A'], 'labels': ['UEFI'], 'masters': []}, 'start': '10240', 'sectors': '217088', 'sectorsize': 512, 'size': '106.00 MB', 'uuid': '5C78-612A', 'holders': []}, 'sda1': {'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1', 'scsi-SQEMU_QEMU_HARDDISK_4fa924fa-33d9-43ce-b208-159d6f6ab539-part1'], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf'], 'labels': ['cloudimg-rootfs'], 'masters': []}, 'start': '2099200', 'sectors': '165672927', 'sectorsize': 512, 'size': '79.00 GB', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf', 'holders': []}}, 'rotational': '1', 'scheduler_mode': 'none', 'sectors': '167772160', 'sectorsize': '512', 'size': '80.00 GB', 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'virtual': 1, 'links': {'ids': [], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '0', 'partitions': {}, 'rotational': '0', 'scheduler_mode': 'none', 'sectors': '0', 'sectorsize': '512', 'size': '0.00 Bytes', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-3', 'value': {'virtual': 1, 'links': {'ids': ['dm-name-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp', 'dm-uuid-CRYPT-LUKS2-958c3d719b3b484b8cbff174ba1f6fac-0bNZRn-Q3R8-1VwJ-TdPW-ktYt-RAGS-sXVFlp'], 'uuids': [], 'labels': [], 'masters': []}, 'vendor': None, 'model': None, 'sas_address': None, 'sas_device_handle': None, 'removable': '0', 'support_discard': '4096', 'partitions': {}, 'rotational': '1', 'scheduler_mode': '', 'sectors': '41902080', 'sectorsize': '512', 'size': '19.98 GB', 'host': '', 'holders': []}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Thursday 26 March 2026  06:12:18 +0000 (0:00:01.383)       1:09:42.306 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Thursday 26 March 2026  06:12:20 +0000 (0:00:01.531)       1:09:43.838 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:12:21 +0000 (0:00:01.150)       1:09:44.988 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:12:22 +0000 (0:00:01.511)       1:09:46.499 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Thursday 26 March 2026  06:12:23 +0000 (0:00:01.109)       1:09:47.609 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Thursday 26 March 2026  06:12:25 +0000 (0:00:01.263)       1:09:48.873 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Thursday 26 March 2026  06:12:26 +0000 (0:00:01.189)       1:09:50.062 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Thursday 26 March 2026  06:12:28 +0000 (0:00:02.086)       1:09:52.149 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Thursday 26 March 2026  06:12:29 +0000 (0:00:01.182)       1:09:53.331 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:12:30 +0000 (0:00:01.303)       1:09:54.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:12:32 +0000 (0:00:01.222)       1:09:55.858 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:12:33 +0000 (0:00:01.130)       1:09:56.988 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:12:34 +0000 (0:00:01.151)       1:09:58.140 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:12:35 +0000 (0:00:01.247)       1:09:59.388 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:12:37 +0000 (0:00:01.441)       1:10:00.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:12:38 +0000 (0:00:01.414)       1:10:02.245 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:12:40 +0000 (0:00:01.423)       1:10:03.669 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:12:41 +0000 (0:00:01.177)       1:10:04.846 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Thursday 26 March 2026  06:12:42 +0000 (0:00:01.339)       1:10:06.186 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Thursday 26 March 2026  06:12:44 +0000 (0:00:02.218)       1:10:08.405 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw when upgrading from stable-3.2] ****************************",
                                "Thursday 26 March 2026  06:12:47 +0000 (0:00:02.296)       1:10:10.701 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 1)] ***************************************************",
                                "Thursday 26 March 2026  06:12:49 +0000 (0:00:02.048)       1:10:12.750 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [Stop ceph rgw (pt. 2)] ***************************************************",
                                "Thursday 26 March 2026  06:12:51 +0000 (0:00:02.507)       1:10:15.258 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  06:12:53 +0000 (0:00:01.940)       1:10:17.198 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  06:12:54 +0000 (0:00:01.092)       1:10:18.291 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  06:12:55 +0000 (0:00:01.144)       1:10:19.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  06:12:56 +0000 (0:00:01.122)       1:10:20.558 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  06:12:58 +0000 (0:00:01.559)       1:10:22.118 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  06:12:59 +0000 (0:00:01.535)       1:10:23.653 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  06:13:01 +0000 (0:00:01.510)       1:10:25.164 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  06:13:02 +0000 (0:00:01.141)       1:10:26.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  06:13:03 +0000 (0:00:01.111)       1:10:27.417 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  06:13:04 +0000 (0:00:01.175)       1:10:28.592 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  06:13:06 +0000 (0:00:01.522)       1:10:30.114 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  06:13:07 +0000 (0:00:01.497)       1:10:31.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  06:13:08 +0000 (0:00:00.751)       1:10:32.363 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  06:13:09 +0000 (0:00:00.756)       1:10:33.120 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  06:13:10 +0000 (0:00:00.820)       1:10:33.941 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  06:13:11 +0000 (0:00:00.831)       1:10:34.772 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  06:13:11 +0000 (0:00:00.808)       1:10:35.581 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  06:13:12 +0000 (0:00:00.800)       1:10:36.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  06:13:13 +0000 (0:00:00.804)       1:10:37.186 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  06:13:14 +0000 (0:00:00.752)       1:10:37.939 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  06:13:15 +0000 (0:00:00.764)       1:10:38.704 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_repository.yml] **************************",
                                "Thursday 26 March 2026  06:13:16 +0000 (0:00:00.975)       1:10:39.679 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_redhat_packages.yml] **************",
                                "Thursday 26 March 2026  06:13:16 +0000 (0:00:00.792)       1:10:40.471 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_suse_packages.yml] ****************",
                                "Thursday 26 March 2026  06:13:17 +0000 (0:00:00.804)       1:10:41.275 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include installs/install_on_debian.yml] ********************",
                                "Thursday 26 March 2026  06:13:18 +0000 (0:00:00.793)       1:10:42.069 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include_tasks installs/install_on_clear.yml] ***************",
                                "Thursday 26 March 2026  06:13:19 +0000 (0:00:00.810)       1:10:42.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Get ceph version] ******************************************",
                                "Thursday 26 March 2026  06:13:20 +0000 (0:00:00.781)       1:10:43.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_version] *************************************",
                                "Thursday 26 March 2026  06:13:20 +0000 (0:00:00.811)       1:10:44.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Set_fact ceph_release - override ceph_release with ceph_stable_release] ***",
                                "Thursday 26 March 2026  06:13:21 +0000 (0:00:00.763)       1:10:45.236 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include create_rbd_client_dir.yml] *************************",
                                "Thursday 26 March 2026  06:13:22 +0000 (0:00:00.778)       1:10:46.015 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_cluster_name.yml] ************************",
                                "Thursday 26 March 2026  06:13:23 +0000 (0:00:00.780)       1:10:46.796 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include configure_memory_allocator.yml] ********************",
                                "Thursday 26 March 2026  06:13:23 +0000 (0:00:00.780)       1:10:47.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-common : Include selinux.yml] ***************************************",
                                "Thursday 26 March 2026  06:13:24 +0000 (0:00:00.779)       1:10:48.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Thursday 26 March 2026  06:13:25 +0000 (0:00:00.797)       1:10:49.153 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Thursday 26 March 2026  06:13:27 +0000 (0:00:01.669)       1:10:50.822 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Thursday 26 March 2026  06:13:29 +0000 (0:00:01.902)       1:10:52.724 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Thursday 26 March 2026  06:13:30 +0000 (0:00:01.127)       1:10:53.852 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Thursday 26 March 2026  06:13:31 +0000 (0:00:01.227)       1:10:55.079 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Thursday 26 March 2026  06:13:32 +0000 (0:00:01.140)       1:10:56.220 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Thursday 26 March 2026  06:13:34 +0000 (0:00:01.774)       1:10:57.995 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Thursday 26 March 2026  06:13:35 +0000 (0:00:01.489)       1:10:59.484 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Thursday 26 March 2026  06:13:36 +0000 (0:00:01.121)       1:11:00.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Thursday 26 March 2026  06:13:37 +0000 (0:00:00.823)       1:11:01.429 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Thursday 26 March 2026  06:13:38 +0000 (0:00:00.769)       1:11:02.198 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Thursday 26 March 2026  06:13:39 +0000 (0:00:01.153)       1:11:03.352 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Thursday 26 March 2026  06:13:41 +0000 (0:00:01.901)       1:11:05.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Thursday 26 March 2026  06:13:42 +0000 (0:00:01.154)       1:11:06.409 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Thursday 26 March 2026  06:13:43 +0000 (0:00:01.154)       1:11:07.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Thursday 26 March 2026  06:13:45 +0000 (0:00:01.164)       1:11:08.727 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Thursday 26 March 2026  06:13:46 +0000 (0:00:01.225)       1:11:09.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Thursday 26 March 2026  06:13:47 +0000 (0:00:01.149)       1:11:11.103 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Thursday 26 March 2026  06:13:48 +0000 (0:00:00.786)       1:11:11.889 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Thursday 26 March 2026  06:13:50 +0000 (0:00:02.128)       1:11:14.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Thursday 26 March 2026  06:13:51 +0000 (0:00:00.789)       1:11:14.807 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Thursday 26 March 2026  06:13:52 +0000 (0:00:01.153)       1:11:15.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Thursday 26 March 2026  06:13:53 +0000 (0:00:01.139)       1:11:17.100 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Thursday 26 March 2026  06:13:54 +0000 (0:00:01.133)       1:11:18.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Thursday 26 March 2026  06:13:55 +0000 (0:00:01.304)       1:11:19.538 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Thursday 26 March 2026  06:13:57 +0000 (0:00:01.134)       1:11:20.673 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Thursday 26 March 2026  06:13:58 +0000 (0:00:01.132)       1:11:21.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Thursday 26 March 2026  06:13:59 +0000 (0:00:01.218)       1:11:23.023 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Thursday 26 March 2026  06:14:00 +0000 (0:00:01.210)       1:11:24.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Thursday 26 March 2026  06:14:01 +0000 (0:00:01.177)       1:11:25.412 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Thursday 26 March 2026  06:14:02 +0000 (0:00:00.832)       1:11:26.244 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Thursday 26 March 2026  06:14:03 +0000 (0:00:01.172)       1:11:27.416 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Thursday 26 March 2026  06:14:09 +0000 (0:00:06.082)       1:11:33.499 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Thursday 26 March 2026  06:14:10 +0000 (0:00:01.095)       1:11:34.594 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Thursday 26 March 2026  06:14:12 +0000 (0:00:01.634)       1:11:36.229 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Thursday 26 March 2026  06:14:14 +0000 (0:00:01.585)       1:11:37.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Thursday 26 March 2026  06:14:14 +0000 (0:00:00.830)       1:11:38.645 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Thursday 26 March 2026  06:14:15 +0000 (0:00:00.800)       1:11:39.446 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Thursday 26 March 2026  06:14:16 +0000 (0:00:00.870)       1:11:40.317 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Thursday 26 March 2026  06:14:17 +0000 (0:00:00.774)       1:11:41.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Thursday 26 March 2026  06:14:18 +0000 (0:00:00.809)       1:11:41.901 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Thursday 26 March 2026  06:14:19 +0000 (0:00:00.775)       1:11:42.677 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Thursday 26 March 2026  06:14:19 +0000 (0:00:00.779)       1:11:43.457 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Thursday 26 March 2026  06:14:20 +0000 (0:00:00.818)       1:11:44.275 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Thursday 26 March 2026  06:14:21 +0000 (0:00:00.830)       1:11:45.105 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Thursday 26 March 2026  06:14:22 +0000 (0:00:00.794)       1:11:45.900 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Thursday 26 March 2026  06:14:23 +0000 (0:00:00.788)       1:11:46.689 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Thursday 26 March 2026  06:14:27 +0000 (0:00:04.177)       1:11:50.866 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Thursday 26 March 2026  06:14:28 +0000 (0:00:00.932)       1:11:51.798 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Thursday 26 March 2026  06:14:32 +0000 (0:00:04.731)       1:11:56.530 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Thursday 26 March 2026  06:14:33 +0000 (0:00:00.778)       1:11:57.308 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Thursday 26 March 2026  06:14:34 +0000 (0:00:00.838)       1:11:58.147 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Thursday 26 March 2026  06:14:35 +0000 (0:00:00.809)       1:11:58.957 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Thursday 26 March 2026  06:14:36 +0000 (0:00:00.825)       1:11:59.783 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Thursday 26 March 2026  06:14:36 +0000 (0:00:00.830)       1:12:00.613 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Thursday 26 March 2026  06:14:37 +0000 (0:00:00.908)       1:12:01.521 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Thursday 26 March 2026  06:14:38 +0000 (0:00:01.113)       1:12:02.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Thursday 26 March 2026  06:14:40 +0000 (0:00:01.068)       1:12:03.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Thursday 26 March 2026  06:14:41 +0000 (0:00:01.058)       1:12:04.762 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Thursday 26 March 2026  06:14:41 +0000 (0:00:00.822)       1:12:05.584 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Thursday 26 March 2026  06:14:43 +0000 (0:00:01.579)       1:12:07.164 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Thursday 26 March 2026  06:14:44 +0000 (0:00:01.420)       1:12:08.584 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:14:46 +0000 (0:00:01.146)       1:12:09.731 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:14:49 +0000 (0:00:03.253)       1:12:12.984 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Thursday 26 March 2026  06:14:51 +0000 (0:00:01.923)       1:12:14.907 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Thursday 26 March 2026  06:14:52 +0000 (0:00:00.832)       1:12:15.740 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Thursday 26 March 2026  06:14:53 +0000 (0:00:01.128)       1:12:16.868 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Thursday 26 March 2026  06:14:54 +0000 (0:00:01.593)       1:12:18.462 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Thursday 26 March 2026  06:15:00 +0000 (0:00:05.203)       1:12:23.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Thursday 26 March 2026  06:15:03 +0000 (0:00:03.118)       1:12:26.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Thursday 26 March 2026  06:15:04 +0000 (0:00:01.764)       1:12:28.548 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Thursday 26 March 2026  06:15:06 +0000 (0:00:01.134)       1:12:29.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Thursday 26 March 2026  06:15:07 +0000 (0:00:01.604)       1:12:31.287 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Thursday 26 March 2026  06:15:09 +0000 (0:00:01.619)       1:12:32.906 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Thursday 26 March 2026  06:15:39 +0000 (0:00:30.561)       1:13:03.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Thursday 26 March 2026  06:15:40 +0000 (0:00:00.768)       1:13:04.237 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Thursday 26 March 2026  06:15:41 +0000 (0:00:00.775)       1:13:05.012 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Thursday 26 March 2026  06:15:42 +0000 (0:00:01.115)       1:13:06.128 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Thursday 26 March 2026  06:15:43 +0000 (0:00:01.109)       1:13:07.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Thursday 26 March 2026  06:15:45 +0000 (0:00:02.032)       1:13:09.270 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Thursday 26 March 2026  06:15:47 +0000 (0:00:02.034)       1:13:11.304 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Thursday 26 March 2026  06:15:49 +0000 (0:00:02.238)       1:13:13.543 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph rbd mirror node] ********************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph nfs node] ***************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph client node] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph-crash daemons] **********************************************",
                                "",
                                "TASK [Stop the ceph-crash service] *********************************************",
                                "Thursday 26 March 2026  06:15:54 +0000 (0:00:04.262)       1:13:17.806 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Mask and disable the ceph-crash service] *********************************",
                                "Thursday 26 March 2026  06:15:56 +0000 (0:00:02.563)       1:13:20.369 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:16:00 +0000 (0:00:03.343)       1:13:23.713 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:16:02 +0000 (0:00:02.461)       1:13:26.174 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Thursday 26 March 2026  06:16:04 +0000 (0:00:02.407)       1:13:28.582 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Thursday 26 March 2026  06:16:07 +0000 (0:00:02.093)       1:13:30.676 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Thursday 26 March 2026  06:16:09 +0000 (0:00:02.227)       1:13:32.903 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Thursday 26 March 2026  06:16:11 +0000 (0:00:02.322)       1:13:35.226 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Thursday 26 March 2026  06:16:13 +0000 (0:00:02.097)       1:13:37.323 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Thursday 26 March 2026  06:16:16 +0000 (0:00:02.526)       1:13:39.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Thursday 26 March 2026  06:16:18 +0000 (0:00:02.109)       1:13:41.960 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Thursday 26 March 2026  06:16:20 +0000 (0:00:02.500)       1:13:44.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Thursday 26 March 2026  06:16:22 +0000 (0:00:01.845)       1:13:46.307 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Thursday 26 March 2026  06:16:24 +0000 (0:00:01.755)       1:13:48.062 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Thursday 26 March 2026  06:16:26 +0000 (0:00:02.545)       1:13:50.608 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Thursday 26 March 2026  06:16:28 +0000 (0:00:02.034)       1:13:52.643 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Thursday 26 March 2026  06:16:31 +0000 (0:00:02.197)       1:13:54.841 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Thursday 26 March 2026  06:16:32 +0000 (0:00:01.813)       1:13:56.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Thursday 26 March 2026  06:16:35 +0000 (0:00:02.039)       1:13:58.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Thursday 26 March 2026  06:16:36 +0000 (0:00:01.825)       1:14:00.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Thursday 26 March 2026  06:16:38 +0000 (0:00:02.062)       1:14:02.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Thursday 26 March 2026  06:16:40 +0000 (0:00:01.830)       1:14:04.412 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Thursday 26 March 2026  06:16:42 +0000 (0:00:02.115)       1:14:06.528 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Thursday 26 March 2026  06:16:44 +0000 (0:00:01.837)       1:14:08.365 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Thursday 26 March 2026  06:16:46 +0000 (0:00:01.850)       1:14:10.215 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Thursday 26 March 2026  06:16:48 +0000 (0:00:02.318)       1:14:12.533 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Thursday 26 March 2026  06:16:51 +0000 (0:00:03.045)       1:14:15.579 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Thursday 26 March 2026  06:16:55 +0000 (0:00:03.509)       1:14:19.088 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Thursday 26 March 2026  06:16:57 +0000 (0:00:02.520)       1:14:21.608 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Thursday 26 March 2026  06:17:00 +0000 (0:00:02.064)       1:14:23.672 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Thursday 26 March 2026  06:17:02 +0000 (0:00:02.694)       1:14:26.367 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Thursday 26 March 2026  06:17:05 +0000 (0:00:02.909)       1:14:29.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Complete upgrade] ********************************************************",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:17:10 +0000 (0:00:04.816)       1:14:34.093 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:17:12 +0000 (0:00:01.712)       1:14:35.806 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Container | disallow pre-reef OSDs and enable all new reef-only functionality] ***",
                                "Thursday 26 March 2026  06:17:13 +0000 (0:00:01.604)       1:14:37.410 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Non container | disallow pre-reef OSDs and enable all new reef-only functionality] ***",
                                "Thursday 26 March 2026  06:17:16 +0000 (0:00:02.350)       1:14:39.761 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Upgrade node-exporter] ***************************************************",
                                "",
                                "TASK [Stop node-exporter] ******************************************************",
                                "Thursday 26 March 2026  06:17:18 +0000 (0:00:01.941)       1:14:41.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  06:17:20 +0000 (0:00:02.558)       1:14:44.261 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-engine : Include pre_requisites/prerequisites.yml] ********",
                                "Thursday 26 March 2026  06:17:23 +0000 (0:00:02.581)       1:14:46.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Container registry authentication] ***************",
                                "Thursday 26 March 2026  06:17:25 +0000 (0:00:02.498)       1:14:49.341 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-node-exporter : Include setup_container.yml] ************************",
                                "Thursday 26 March 2026  06:17:28 +0000 (0:00:02.578)       1:14:51.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Upgrade monitoring node] *************************************************",
                                "",
                                "TASK [Stop monitoring services] ************************************************",
                                "Thursday 26 March 2026  06:17:31 +0000 (0:00:02.982)       1:14:54.903 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=alertmanager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=prometheus) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=grafana-server) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addr fact - ipv4] ************************",
                                "Thursday 26 March 2026  06:17:32 +0000 (0:00:01.282)       1:14:56.185 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addr fact - ipv6] ************************",
                                "Thursday 26 March 2026  06:17:33 +0000 (0:00:01.158)       1:14:57.344 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addrs fact - ipv4] ***********************",
                                "Thursday 26 March 2026  06:17:34 +0000 (0:00:01.175)       1:14:58.520 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addrs fact - ipv6] ***********************",
                                "Thursday 26 March 2026  06:17:35 +0000 (0:00:01.124)       1:14:59.645 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Create prometheus directories] *************************",
                                "Thursday 26 March 2026  06:17:37 +0000 (0:00:01.176)       1:15:00.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/prometheus) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/var/lib/prometheus) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Write prometheus config file] **************************",
                                "Thursday 26 March 2026  06:17:38 +0000 (0:00:01.121)       1:15:01.943 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Make sure the alerting rules directory exists] *********",
                                "Thursday 26 March 2026  06:17:39 +0000 (0:00:01.148)       1:15:03.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Copy alerting rules] ***********************************",
                                "Thursday 26 March 2026  06:17:40 +0000 (0:00:01.112)       1:15:04.204 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Create alertmanager directories] ***********************",
                                "Thursday 26 March 2026  06:17:41 +0000 (0:00:01.343)       1:15:05.548 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/alertmanager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/var/lib/alertmanager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Write alertmanager config file] ************************",
                                "Thursday 26 March 2026  06:17:43 +0000 (0:00:01.136)       1:15:06.684 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-prometheus : Include setup_container.yml] ***************************",
                                "Thursday 26 March 2026  06:17:44 +0000 (0:00:01.103)       1:15:07.788 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-grafana : Include setup_container.yml] ******************************",
                                "Thursday 26 March 2026  06:17:45 +0000 (0:00:01.103)       1:15:08.891 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [ceph-grafana : Include configure_grafana.yml] ****************************",
                                "Thursday 26 March 2026  06:17:46 +0000 (0:00:01.116)       1:15:10.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Upgrade ceph dashboard] **************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Thursday 26 March 2026  06:17:47 +0000 (0:00:01.641)       1:15:11.649 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addr fact - ipv4] ************************",
                                "Thursday 26 March 2026  06:17:49 +0000 (0:00:01.793)       1:15:13.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addr fact - ipv6] ************************",
                                "Thursday 26 March 2026  06:17:51 +0000 (0:00:01.527)       1:15:14.970 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addrs fact - ipv4] ***********************",
                                "Thursday 26 March 2026  06:17:52 +0000 (0:00:01.430)       1:15:16.401 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set grafana_server_addrs fact - ipv6] ***********************",
                                "Thursday 26 March 2026  06:17:54 +0000 (0:00:01.759)       1:15:18.160 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-dashboard : Include configure_dashboard.yml] ************************",
                                "Thursday 26 March 2026  06:17:55 +0000 (0:00:01.437)       1:15:19.597 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-dashboard : Print dashboard URL] ************************************",
                                "Thursday 26 March 2026  06:17:57 +0000 (0:00:01.397)       1:15:20.995 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Switch any existing crush buckets to straw2] *****************************",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Thursday 26 March 2026  06:17:59 +0000 (0:00:02.089)       1:15:23.085 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Thursday 26 March 2026  06:18:00 +0000 (0:00:01.526)       1:15:24.611 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set_fact ceph_cmd] *******************************************************",
                                "Thursday 26 March 2026  06:18:02 +0000 (0:00:01.243)       1:15:25.854 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Backup the crushmap] *****************************************************",
                                "Thursday 26 March 2026  06:18:03 +0000 (0:00:01.114)       1:15:26.969 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Switch crush buckets to straw2] ******************************************",
                                "Thursday 26 March 2026  06:18:06 +0000 (0:00:02.960)       1:15:29.929 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Remove crushmap backup] **************************************************",
                                "Thursday 26 March 2026  06:18:09 +0000 (0:00:03.295)       1:15:33.225 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Show ceph status] ********************************************************",
                                "",
                                "TASK [Set_fact container_exec_cmd_status] **************************************",
                                "Thursday 26 March 2026  06:18:11 +0000 (0:00:01.963)       1:15:35.189 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Show ceph status] ********************************************************",
                                "Thursday 26 March 2026  06:18:13 +0000 (0:00:02.269)       1:15:37.458 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Show all daemons version] ************************************************",
                                "Thursday 26 March 2026  06:18:16 +0000 (0:00:02.338)       1:15:39.797 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : ok=0    changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=76  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=248 \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=376 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=191 \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=350 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=196 \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=351 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=317 \u001b[0m \u001b[0;33mchanged=21  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=362 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=307 \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=359 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=303 \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=344 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  06:18:19 +0000 (0:00:03.761)       1:15:43.559 ******** ",
                                "=============================================================================== ",
                                "Disable pg autoscale on pools ------------------------------------------ 74.54s",
                                "Re-enable pg autoscale on pools ---------------------------------------- 74.51s",
                                "Gather and delegate facts ---------------------------------------------- 31.50s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 31.45s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 31.26s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 30.56s",
                                "Waiting for clean pgs... ----------------------------------------------- 29.82s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 28.66s",
                                "ceph-config : Set config to cluster ------------------------------------ 23.07s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 23.01s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 22.82s",
                                "Stop ceph mgr ---------------------------------------------------------- 17.87s",
                                "Create potentially missing keys (rbd and rbd-mirror) ------------------- 16.74s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 15.47s",
                                "ceph-config : Set config to cluster ------------------------------------ 14.42s",
                                "ceph-config : Set osd_memory_target to cluster host config ------------- 12.82s",
                                "ceph-config : Set osd_memory_target to cluster host config ------------- 12.81s",
                                "Stop ceph osd ---------------------------------------------------------- 12.80s",
                                "Stop standby ceph mds -------------------------------------------------- 11.17s",
                                "Stop ceph mon ---------------------------------------------------------- 10.37s",
                                "+ osism apply cephclient",
                                "\u001b[32m2026-03-26 06:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21ea1124-23b4-48c3-93da-984335f8d483 (cephclient) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-26 06:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 21ea1124-23b4-48c3-93da-984335f8d483 (cephclient) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_play_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x79be164d1be0>): Expecting value: line 2 column 1 (char 1)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Failure using method (v2_playbook_on_task_start) in callback plugin\u001b[0m",
                                "\u001b[1;35m(<ansible.plugins.callback.ara_default.CallbackModule object at\u001b[0m",
                                "\u001b[1;35m0x79be164d1be0>): 'NoneType' object is not subscriptable\u001b[0m",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Thursday 26 March 2026  06:18:31 +0000 (0:00:02.472)       0:00:02.472 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Thursday 26 March 2026  06:18:32 +0000 (0:00:00.801)       0:00:03.273 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Thursday 26 March 2026  06:18:33 +0000 (0:00:01.721)       0:00:04.995 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Thursday 26 March 2026  06:18:34 +0000 (0:00:01.120)       0:00:06.116 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Thursday 26 March 2026  06:18:35 +0000 (0:00:00.932)       0:00:07.049 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Thursday 26 March 2026  06:18:36 +0000 (0:00:00.951)       0:00:08.001 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Thursday 26 March 2026  06:18:38 +0000 (0:00:01.275)       0:00:09.276 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Thursday 26 March 2026  06:18:42 +0000 (0:00:04.160)       0:00:13.437 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Thursday 26 March 2026  06:18:42 +0000 (0:00:00.508)       0:00:13.946 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Thursday 26 March 2026  06:18:42 +0000 (0:00:00.152)       0:00:14.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=8   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Thursday 26 March 2026  06:18:44 +0000 (0:00:01.113)       0:00:15.211 ******** ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 4.16s",
                                "osism.services.cephclient : Create required directories ----------------- 1.72s",
                                "osism.services.cephclient : Manage cephclient service ------------------- 1.28s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.12s",
                                "osism.services.cephclient : Include rook task --------------------------- 1.11s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.95s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.93s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.80s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.15s",
                                "+ [[ false == \\f\\a\\l\\s\\e ]]",
                                "+ sh -c /opt/configuration/scripts/upgrade/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=9.5.0",
                                "++ MANAGER_VERSION=9.5.0",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=false",
                                "++ TEMPEST=false",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.54",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "++ export RABBITMQ3TO4=true",
                                "++ RABBITMQ3TO4=true",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "++ export MANAGER_VERSION=10.0.0-rc.1",
                                "++ MANAGER_VERSION=10.0.0-rc.1",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ osism migrate rabbitmq3to4 prepare",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to RabbitMQ Management API at 192.168.16.10:15672 (node: testbed-node-0) as openstack...\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mKolla configuration check passed\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreated vhost 'openstack' with default_queue_type=quorum\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSet permissions for user 'openstack' on vhost 'openstack'\u001b[0m",
                                "+ osism migrate rabbitmq3to4 list",
                                "\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to RabbitMQ Management API at 192.168.16.10:15672 (node: testbed-node-0) as openstack...\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFound 208 classic queue(s) in vhost '/':\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - alarm.all.sample (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - alarming.sample (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers.barbican.queue (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_b4e210291c1f4595a8110eea41f3b718 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_e5817fbc18e24736ae19aea2dd46443c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican.workers_fanout_fb5c5d01a8bc44969546e21abf2484cc (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - barbican_notifications.info (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_35733cd31394488d9e3db0d42e40e34c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_3c6a3e144f2243b2bc8f0af87b8fdf99 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_3e97b3e6bc4c46eab56fa9a43109ebd7 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_6bd3253d59d0488c8ecbab32dee32db2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_82cf3e0940194386804159a753c7460c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - central_fanout_d4f9292219c84761a99423dd67f1398c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_8f5f559e775243a98c2c1f750404a4f1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_bf6b6bf9a62148dea1244fab3e97b501 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-backup_fanout_f2d658f5fc36455a8931e7580182289c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_7eb0367bc74f44d2b4625e36bc5d642d (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_d047f442bc244458a150b5609e4af3ec (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-scheduler_fanout_f872bbe8a39c496c82652ebaf402b14a (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-0@rbd-volumes_fanout_a6533171c6994adf9bcce0cd216bf57c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-1@rbd-volumes_fanout_b4bcfca53a99458ab09955313468afc6 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume.testbed-node-2@rbd-volumes_fanout_ade417b8c94e4319896a910815b50d48 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_029b93b2eb2647bebe3b03dbcaf16cff (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_8c9dd967d35c4794ac8b90812623d27b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - cinder-volume_fanout_8fd2c7d0d57d479098d7dc9f2a8a7324 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-3 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-4 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute.testbed-node-5 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_1633d9b6ca8d45e3980c3aa8b6d1479f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_2ed9f7af58614a9fac21fe0ebf645431 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - compute_fanout_dce43fa4b04f4a428d9422d5117ae051 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_0a087e21fbc9400eaca4d0d8c28e18e1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_2779a7fe25ce4c3292b21b1ec783f5a8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_3807aa7d35c84957b24855e76e40df5c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_39ddbc7364ef4e6093e49dc1f29da98e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_3a1d92acabb3468e9eeb34fd1571f679 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - conductor_fanout_434589943aa54e9c8edd5825fd358e9e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - event.sample (vhost: /, messages: 9)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.dxt76xirij65 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.u74qoctmj2at (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor.ylgl7fnvnb3q (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_114f868d641941c393c267ebefb2d881 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_25275d6376434f9a88947423f3f0ddf9 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_4ada99082e6346f38692f45c23a45fb8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_4b77c13e89a54dfdab9272dd2f0ef97f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_a7c5e48d18c443a1956b99c2edbcbc94 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_b3b098cb100849bab05884917ef2e4e2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_e4ecb03c594f47fe9337c88ff65131bd (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_ee984a30032243dfb0cc80ab0c2b8384 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - magnum-conductor_fanout_f4be591b6d034162b990cb9296faba3f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_c0ecdb1bd3924f28b9e1bbc7eccc646b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_f03d3befe0314d879d7f796de934dd9d (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-data_fanout_f85984553bde4d31bf278d156c7ac211 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_a136d7d13b474010a342264c2c907e16 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_e2eb07a671e64e1686089b8a01af098e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-scheduler_fanout_f7be0f8d31c34c4da429bfe2aca47847 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-0@cephfsnative1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-1@cephfsnative1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share.testbed-node-2@cephfsnative1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_7805b1c8e1414693b0ba1f609b02809a (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_bc63684f7fde4834a4efc8c16b403a68 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - manila-share_fanout_fc5e9c2e25184fcfb3dfda7a4dc4dc33 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.audit (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.critical (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.debug (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.error (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.info (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.sample (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - notifications.warn (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_0fac4d064de84e5bbd6d466dd834e1f4 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_86ab57fd165d4bd783e3c86dd75f8847 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - octavia_provisioning_v2_fanout_c60f129848694c27b399ca7ddbf06f06 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_1598d067d87e4d04b00c428b4b47727f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_36061f07216b4217adb49fe71c35eaf8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_43e908b0d0d945869d271173a5152aba (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_ae6a2357e6a34a50b6620a7e1a181a77 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_cbd89788cf2f40d2a554ea5281955327 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - producer_fanout_e0f7d2d7fb274163a86be0a37dfd2cd8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_44a4c0e6ec6844979a126b54345b60f3 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_63aafcca795041abb90dcd32d46b2b55 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_80382bcbeca9481f812885405d457324 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_a296e3d4085d407584dbef3c971a0f2a (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_b4cf746c93324effb7ff69cab9c2350a (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_c3871e367a2d4d3baf01c5aea2c1919a (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f41bb944c69d473ab16fa6b5d6d35e10 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f7d147aad30f4bcb9adb25eeb33c86c0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-plugin_fanout_f9b48b37b74d417d94f841025641bd73 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_1f4456cdaa8043e3be2805ad39d5ab25 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_26775edc33404dbc9abe1e144bd80525 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_31a15587650e47c3bde5abc053e22c01 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_32ec2ee87c0245788dffd816ea3c023c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_4a6c18d1e8814a9dae92e88643a4f6a3 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_538ba911017345b08e67db1a5c5b55b5 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_5897b7a973ad4acc91343aa05c7af3ff (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6319d5e241c541e58310e16e9dc52144 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6b360b1659a14d0da13ea3b346e563aa (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_6f16a59b15654f4dbcf85abee6644a2b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_94ee497291ae489a9c6b238ca80e38fd (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_9e52c6dc9ed24a7b80ea13a94674ad74 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_aa918d06050f4bd7a2ee224ba419ce31 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_c1cdb63a10e04fc6bf41640f72fa8159 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_cfb3d53116bd463396bf620650870a65 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_d09b359918574dcab62f2ee6b5dcdca8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_dad219db8f49450d9a2b986223f703bf (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-reports-plugin_fanout_dd786927624c419ead6015dcfc58770f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_1cb9224fc37a4a48bfacc80953dd403f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_3135493c2b264f71bb9bbad34fb00452 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_3d6dfb328ccc46a6a9cb2f4b96e3d697 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_408591ee4ea44bc4bacb6ad94f1cf98e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_4500faec64a44eedbcb004f0e67b9b20 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_64da208a82c3494385ded2e418c9a122 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_651674cdba1c43a9813d7ecb595a9cfb (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_c6e201d63611428cad0b60e848404955 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - q-server-resource-versions_fanout_fa758238da8947f4b71f8bcdaabb883b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_0143f57daf2b426b89201cef986f95e2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_1607a4c70ae045398f71fc2d451c5a27 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_2d05cc2aedda45ea9c31797c9488552e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_2df21c6aa5e242c2a4bd19ebf039c414 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_42c674747649424c96f46572b03ea0da (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_4ef0c8bf75cf449ca8256621ad7454ac (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_550a4fdad56e49009b36c1374a865c76 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_919bd19646db497b8240f22dec5e8b0c (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_91fe99c6abc64314ada5daa43f18f50b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_9db8e0e8fdb44324a47d8bfc1ac1f280 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a0bf128c59ef4dcf98d30d7684f34f2e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a1d93d6f81c749c9b16200e2c0c20728 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_a363e6378e634bcaaa24df418226ca2b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_d2aaa10c347d4f878af4a93b4462aca8 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_dcf1dff11157461c91e0f07e1149d443 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_e5a2bc26f56f421d9a8b1028dfa744bf (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_e6549370496a4080a6647b33c1437c75 (vhost: /, messages: 1)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_ece0d2b5146749299699600bd6620114 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - reply_ee22625ab8d44350bdd04b376e72d57f (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_17c056e993ef47e3acc112af16d457e3 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_495244cec2c64f3c9bdf3534c4821ed3 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_4f6c87c162d1459ea749bf93d9575641 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_5825e34b1d114c22a0ea70626d5b3c06 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_9dd488cc166144d3b66b6167500e5b69 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - scheduler_fanout_d8d2fe93979a46f4ad96c7e79aad00d7 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-0 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-1 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker.testbed-node-2 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_360ce71e2a2b43c898e071ad25434240 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_4b3c4dcb71b947008464d6f8006593a6 (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_4f0de0701f684613b3655e8683bea71b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_5a8bae1580fd4d94b676dd0da9c1331b (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_c08facaf21194343b22a11cadc24a44e (vhost: /, messages: 0)\u001b[0m",
                                "\u001b[32m2026-03-26 06:19:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m  - worker_fanout_cc85c421d64b4d6b9fbd61b01ede5463 (vhost: /, messages: 0)\u001b[0m",
                                "+ osism migrate rabbitmq3to4 list-exchanges",
                                "usage: osism migrate rabbitmq3to4 [-h] [--server SERVER] [--dry-run]",
                                "                                  [--no-close-connections] [--quorum]",
                                "                                  [--vhost VHOST]",
                                "                                  [{list,delete,prepare,check}]",
                                "                                  [{aodh,barbican,ceilometer,cinder,designate,notifications,manager,magnum,manila,neutron,nova,octavia}]",
                                "osism migrate rabbitmq3to4: error: argument command: invalid choice: 'list-exchanges' (choose from list, delete, prepare, check)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-ba6a-269d-00000000002b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:28.816979Z",
                            "start": "2026-03-26T04:14:59.932110Z"
                        },
                        "id": "0242ac1f-6606-ba6a-269d-00000000002b",
                        "name": "Run upgrade"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 11,
            "failures": 1,
            "ignored": 0,
            "ok": 30,
            "rescued": 0,
            "skipped": 6,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:19:32.694804Z",
                    "start": "2026-03-26T06:19:29.983666Z"
                },
                "id": "0242ac1f-6606-8775-4b7b-000000000002",
                "name": "Post output play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:30.057034Z",
                            "start": "2026-03-26T06:19:30.001546Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000007",
                        "name": "Register sources"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "sudo",
                                "-n",
                                "true"
                            ],
                            "delta": "0:00:00.016472",
                            "end": "2026-03-26 06:19:30.961272",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "sudo -n true",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8775-4b7b-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-03-26 06:19:30.944800",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sudo: a password is required",
                            "stdout_lines": [
                                "sudo: a password is required"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8775-4b7b-000000000009-1-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:31.122780Z",
                            "start": "2026-03-26T06:19:30.109793Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000009",
                        "name": "Check sudo"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:31.162193Z",
                            "start": "2026-03-26T06:19:31.136745Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000000a",
                        "name": "Set source and destination for files and folders"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "all_sources": []
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:31.249932Z",
                            "start": "2026-03-26T06:19:31.177504Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000000c",
                        "name": "Build a list of source, dest dictionaries"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/docs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/docs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "docs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/logs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "logs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.236021Z",
                            "start": "2026-03-26T06:19:31.258955Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000000d",
                        "name": "Ensure target folders exist"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "results": [],
                            "skipped": true,
                            "skipped_reason": "No items in the list"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.275986Z",
                            "start": "2026-03-26T06:19:32.250769Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000000f",
                        "name": "Copy files and folders to staging folder"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed03/zuul-output/logs"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed03/zuul-output/logs"
                                }
                            },
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "dest": "/home/zuul-testbed03/zuul-output/logs",
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": "u=rwX,g=rX,o=rX",
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed03/zuul-output/logs",
                                    "recurse": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed03",
                            "path": "/home/zuul-testbed03/zuul-output/logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.572828Z",
                            "start": "2026-03-26T06:19:32.290519Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000011",
                        "name": "Make all log files readable"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stage_output_renames",
                            "changed": false,
                            "false_condition": "extensions_to_txt is defined and extensions_to_txt is not none",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.617506Z",
                            "start": "2026-03-26T06:19:32.581929Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000012",
                        "name": "Rename log files that match extensions_to_txt"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "find",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.658472Z",
                            "start": "2026-03-26T06:19:32.633643Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000014",
                        "name": "Discover log files for compression"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-8775-4b7b-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.694804Z",
                            "start": "2026-03-26T06:19:32.670718Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000015",
                        "name": "Archive everything from logs"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:19:53.404417Z",
                    "start": "2026-03-26T06:19:32.712634Z"
                },
                "id": "0242ac1f-6606-8775-4b7b-000000000017",
                "name": "Post cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.777734Z",
                            "start": "2026-03-26T06:19:32.721279Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-000000000019",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:32.824604Z",
                            "start": "2026-03-26T06:19:32.789753Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000001a",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:20.185435",
                            "end": "2026-03-26 06:19:53.328965",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8775-4b7b-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 06:19:33.143530",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-26 06:19:33 - clean up servers\n2026-03-26 06:19:34 - testbed-manager\n2026-03-26 06:19:34 - testbed-node-4\n2026-03-26 06:19:34 - testbed-node-5\n2026-03-26 06:19:34 - testbed-node-0\n2026-03-26 06:19:34 - testbed-node-3\n2026-03-26 06:19:34 - testbed-node-1\n2026-03-26 06:19:34 - testbed-node-2\n2026-03-26 06:19:34 - clean up keypairs\n2026-03-26 06:19:34 - testbed\n2026-03-26 06:19:34 - wait for servers to be gone\n2026-03-26 06:19:45 - clean up ports\n2026-03-26 06:19:45 - 0d011944-d664-48dc-afd5-e876547c06cd\n2026-03-26 06:19:46 - 4acb0156-c5c7-4457-a252-e47665f1efbd\n2026-03-26 06:19:46 - 5d571e2f-b701-4263-883f-e7221cc0c90a\n2026-03-26 06:19:46 - be5d018b-5284-4a13-84e4-e85247bdd4a2\n2026-03-26 06:19:47 - c09a035e-b37c-4623-b251-1a605583e2e5\n2026-03-26 06:19:47 - d0917be5-d774-4ae8-aa8e-88d5b51d1da3\n2026-03-26 06:19:47 - da761b17-04fe-4927-b0ae-964e0d48a3d8\n2026-03-26 06:19:47 - clean up volumes\n2026-03-26 06:19:48 - testbed-volume-3-node-base\n2026-03-26 06:19:48 - testbed-volume-2-node-base\n2026-03-26 06:19:48 - testbed-volume-5-node-base\n2026-03-26 06:19:48 - testbed-volume-4-node-base\n2026-03-26 06:19:48 - testbed-volume-manager-base\n2026-03-26 06:19:48 - testbed-volume-1-node-base\n2026-03-26 06:19:48 - testbed-volume-0-node-base\n2026-03-26 06:19:48 - testbed-volume-0-node-3\n2026-03-26 06:19:48 - testbed-volume-4-node-4\n2026-03-26 06:19:48 - testbed-volume-5-node-5\n2026-03-26 06:19:48 - testbed-volume-8-node-5\n2026-03-26 06:19:48 - testbed-volume-3-node-3\n2026-03-26 06:19:48 - testbed-volume-1-node-4\n2026-03-26 06:19:48 - testbed-volume-7-node-4\n2026-03-26 06:19:48 - testbed-volume-2-node-5\n2026-03-26 06:19:48 - testbed-volume-6-node-3\n2026-03-26 06:19:48 - disconnect routers\n2026-03-26 06:19:48 - testbed\n2026-03-26 06:19:50 - clean up subnets\n2026-03-26 06:19:50 - subnet-testbed-management\n2026-03-26 06:19:50 - clean up networks\n2026-03-26 06:19:50 - net-testbed-management\n2026-03-26 06:19:51 - clean up security groups\n2026-03-26 06:19:51 - testbed-management\n2026-03-26 06:19:51 - testbed-node\n2026-03-26 06:19:51 - clean up floating ips\n2026-03-26 06:19:51 - 81.163.192.54\n2026-03-26 06:19:52 - clean up routers\n2026-03-26 06:19:52 - testbed",
                            "stdout_lines": [
                                "2026-03-26 06:19:33 - clean up servers",
                                "2026-03-26 06:19:34 - testbed-manager",
                                "2026-03-26 06:19:34 - testbed-node-4",
                                "2026-03-26 06:19:34 - testbed-node-5",
                                "2026-03-26 06:19:34 - testbed-node-0",
                                "2026-03-26 06:19:34 - testbed-node-3",
                                "2026-03-26 06:19:34 - testbed-node-1",
                                "2026-03-26 06:19:34 - testbed-node-2",
                                "2026-03-26 06:19:34 - clean up keypairs",
                                "2026-03-26 06:19:34 - testbed",
                                "2026-03-26 06:19:34 - wait for servers to be gone",
                                "2026-03-26 06:19:45 - clean up ports",
                                "2026-03-26 06:19:45 - 0d011944-d664-48dc-afd5-e876547c06cd",
                                "2026-03-26 06:19:46 - 4acb0156-c5c7-4457-a252-e47665f1efbd",
                                "2026-03-26 06:19:46 - 5d571e2f-b701-4263-883f-e7221cc0c90a",
                                "2026-03-26 06:19:46 - be5d018b-5284-4a13-84e4-e85247bdd4a2",
                                "2026-03-26 06:19:47 - c09a035e-b37c-4623-b251-1a605583e2e5",
                                "2026-03-26 06:19:47 - d0917be5-d774-4ae8-aa8e-88d5b51d1da3",
                                "2026-03-26 06:19:47 - da761b17-04fe-4927-b0ae-964e0d48a3d8",
                                "2026-03-26 06:19:47 - clean up volumes",
                                "2026-03-26 06:19:48 - testbed-volume-3-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-2-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-5-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-4-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-manager-base",
                                "2026-03-26 06:19:48 - testbed-volume-1-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-0-node-base",
                                "2026-03-26 06:19:48 - testbed-volume-0-node-3",
                                "2026-03-26 06:19:48 - testbed-volume-4-node-4",
                                "2026-03-26 06:19:48 - testbed-volume-5-node-5",
                                "2026-03-26 06:19:48 - testbed-volume-8-node-5",
                                "2026-03-26 06:19:48 - testbed-volume-3-node-3",
                                "2026-03-26 06:19:48 - testbed-volume-1-node-4",
                                "2026-03-26 06:19:48 - testbed-volume-7-node-4",
                                "2026-03-26 06:19:48 - testbed-volume-2-node-5",
                                "2026-03-26 06:19:48 - testbed-volume-6-node-3",
                                "2026-03-26 06:19:48 - disconnect routers",
                                "2026-03-26 06:19:48 - testbed",
                                "2026-03-26 06:19:50 - clean up subnets",
                                "2026-03-26 06:19:50 - subnet-testbed-management",
                                "2026-03-26 06:19:50 - clean up networks",
                                "2026-03-26 06:19:50 - net-testbed-management",
                                "2026-03-26 06:19:51 - clean up security groups",
                                "2026-03-26 06:19:51 - testbed-management",
                                "2026-03-26 06:19:51 - testbed-node",
                                "2026-03-26 06:19:51 - clean up floating ips",
                                "2026-03-26 06:19:51 - 81.163.192.54",
                                "2026-03-26 06:19:52 - clean up routers",
                                "2026-03-26 06:19:52 - testbed"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8775-4b7b-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:53.404417Z",
                            "start": "2026-03-26T06:19:32.866128Z"
                        },
                        "id": "0242ac1f-6606-8775-4b7b-00000000001b",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 7,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/cleanup.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:19:56.983016Z",
                    "start": "2026-03-26T06:19:54.282531Z"
                },
                "id": "0242ac1f-6606-a26a-fbbb-000000000002",
                "name": "Cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:54.367363Z",
                            "start": "2026-03-26T06:19:54.298518Z"
                        },
                        "id": "0242ac1f-6606-a26a-fbbb-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:54.411173Z",
                            "start": "2026-03-26T06:19:54.376206Z"
                        },
                        "id": "0242ac1f-6606-a26a-fbbb-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.349109",
                            "end": "2026-03-26 06:19:56.620464",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-a26a-fbbb-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-26 06:19:55.271355",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-26 06:19:55 - clean up servers\n2026-03-26 06:19:56 - clean up keypairs\n2026-03-26 06:19:56 - wait for servers to be gone\n2026-03-26 06:19:56 - clean up ports\n2026-03-26 06:19:56 - clean up volumes\n2026-03-26 06:19:56 - disconnect routers\n2026-03-26 06:19:56 - clean up subnets\n2026-03-26 06:19:56 - clean up networks\n2026-03-26 06:19:56 - clean up security groups\n2026-03-26 06:19:56 - clean up floating ips\n2026-03-26 06:19:56 - clean up routers",
                            "stdout_lines": [
                                "2026-03-26 06:19:55 - clean up servers",
                                "2026-03-26 06:19:56 - clean up keypairs",
                                "2026-03-26 06:19:56 - wait for servers to be gone",
                                "2026-03-26 06:19:56 - clean up ports",
                                "2026-03-26 06:19:56 - clean up volumes",
                                "2026-03-26 06:19:56 - disconnect routers",
                                "2026-03-26 06:19:56 - clean up subnets",
                                "2026-03-26 06:19:56 - clean up networks",
                                "2026-03-26 06:19:56 - clean up security groups",
                                "2026-03-26 06:19:56 - clean up floating ips",
                                "2026-03-26 06:19:56 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-a26a-fbbb-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:56.983016Z",
                            "start": "2026-03-26T06:19:54.470242Z"
                        },
                        "id": "0242ac1f-6606-a26a-fbbb-000000000006",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "2",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post-fetch.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:20:01.585236Z",
                    "start": "2026-03-26T06:19:57.845351Z"
                },
                "id": "0242ac1f-6606-3ee1-631e-000000000002",
                "name": "Base post-fetch"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-3ee1-631e-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:57.918047Z",
                            "start": "2026-03-26T06:19:57.861216Z"
                        },
                        "id": "0242ac1f-6606-3ee1-631e-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-3ee1-631e-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:57.978498Z",
                            "start": "2026-03-26T06:19:57.932348Z"
                        },
                        "id": "0242ac1f-6606-3ee1-631e-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-3ee1-631e-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:19:59.081325Z",
                            "start": "2026-03-26T06:19:57.984432Z"
                        },
                        "id": "0242ac1f-6606-3ee1-631e-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/logs/ /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/artifacts/ /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/docs/ /var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-3ee1-631e-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:20:01.514758Z",
                            "start": "2026-03-26T06:19:59.097371Z"
                        },
                        "id": "0242ac1f-6606-3ee1-631e-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "msg": "All items skipped",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "docs"
                                }
                            ],
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-3ee1-631e-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:20:01.585236Z",
                            "start": "2026-03-26T06:20:01.536743Z"
                        },
                        "id": "0242ac1f-6606-3ee1-631e-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 3,
            "rescued": 0,
            "skipped": 2,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "3",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-26T06:20:03.467442Z",
                    "start": "2026-03-26T06:20:02.474780Z"
                },
                "id": "0242ac1f-6606-e566-987e-000000000002",
                "name": "Base post"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiGosbjFN+VKLVAAW4MaKK1p284N8RWMRKkKqiULio+JY+rR819U6DjxCGg+fXIBh3XGz2z3gDc6WfZgG2AKcQKctNOf+fGww3c6RC6SxoXeKXZsO1Cu1c3iNH3xmavhWZgel4BbM7u1a3rTS/rs0XIYIEynby207yXEYURr13Z+0qBp3SH5VIadwoHjnalGWqHXcUDBnnnpSZacsS1s5IeDjmZabYyTq3bV7U/bdMo6kWZ2Lk9gmzqdFKgHW26n6AGTbqLPAGxWa6VpPFqNxmAxgP9wKDIHqPbjTTSjrgWsWBgNg91bYRoUm6nwmu7x2kBkRrfCP43MeylLqZeuurnqYiFOK3OEkzi+N4+PnCqE3F3hF/POX9Mu1tjjSZEffumApnUzB/h/6T9KdIr8maOoGP38ZyUko+vHcCnogNY3zeoidk9b3uVMlbPxeymXtTn5dV2UYXaXAwkybq1MGO+rNpk+hUTFrV61Ym3Pq/7ZJyCYSAkcbouuQ6eRHhNzs= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul-testbed03",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiGosbjFN+VKLVAAW4MaKK1p284N8RWMRKkKqiULio+JY+rR819U6DjxCGg+fXIBh3XGz2z3gDc6WfZgG2AKcQKctNOf+fGww3c6RC6SxoXeKXZsO1Cu1c3iNH3xmavhWZgel4BbM7u1a3rTS/rs0XIYIEynby207yXEYURr13Z+0qBp3SH5VIadwoHjnalGWqHXcUDBnnnpSZacsS1s5IeDjmZabYyTq3bV7U/bdMo6kWZ2Lk9gmzqdFKgHW26n6AGTbqLPAGxWa6VpPFqNxmAxgP9wKDIHqPbjTTSjrgWsWBgNg91bYRoUm6nwmu7x2kBkRrfCP43MeylLqZeuurnqYiFOK3OEkzi+N4+PnCqE3F3hF/POX9Mu1tjjSZEffumApnUzB/h/6T9KdIr8maOoGP38ZyUko+vHcCnogNY3zeoidk9b3uVMlbPxeymXtTn5dV2UYXaXAwkybq1MGO+rNpk+hUTFrV61Ym3Pq/7ZJyCYSAkcbouuQ6eRHhNzs= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul-testbed03",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-e566-987e-000000000005",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/6d507829b6994532b2cddf15505f7f09/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-26T06:20:03.467442Z",
                            "start": "2026-03-26T06:20:02.489526Z"
                        },
                        "id": "0242ac1f-6606-e566-987e-000000000007",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
